flush_list 466 drivers/md/dm-log-userspace-base.c static int flush_one_by_one(struct log_c *lc, struct list_head *flush_list) flush_list 471 drivers/md/dm-log-userspace-base.c list_for_each_entry(fe, flush_list, list) { flush_list 483 drivers/md/dm-log-userspace-base.c static int flush_by_group(struct log_c *lc, struct list_head *flush_list, flush_list 496 drivers/md/dm-log-userspace-base.c while (!list_empty(flush_list)) { flush_list 499 drivers/md/dm-log-userspace-base.c list_for_each_entry_safe(fe, tmp_fe, flush_list, list) { flush_list 529 drivers/md/dm-log-userspace-base.c list_splice_init(&tmp_list, flush_list); flush_list 530 drivers/md/dm-log-userspace-base.c r = flush_one_by_one(lc, flush_list); flush_list 540 drivers/md/dm-log-userspace-base.c list_splice_init(&tmp_list, flush_list); flush_list 176 drivers/md/dm-writecache.c struct bio_list flush_list; flush_list 1110 drivers/md/dm-writecache.c bio = bio_list_pop(&wc->flush_list); flush_list 1144 drivers/md/dm-writecache.c if (bio_list_empty(&wc->flush_list)) flush_list 1146 drivers/md/dm-writecache.c bio_list_add(&wc->flush_list, bio); flush_list 2114 drivers/md/dm-writecache.c bio_list_init(&wc->flush_list); flush_list 64 drivers/md/dm-zoned-target.c struct bio_list flush_list; flush_list 510 drivers/md/dm-zoned-target.c bio = bio_list_pop(&dmz->flush_list); flush_list 655 drivers/md/dm-zoned-target.c bio_list_add(&dmz->flush_list, bio); flush_list 824 drivers/md/dm-zoned-target.c bio_list_init(&dmz->flush_list); flush_list 75 kernel/bpf/cpumap.c struct list_head __percpu *flush_list; flush_list 118 kernel/bpf/cpumap.c cmap->flush_list = alloc_percpu(struct list_head); flush_list 119 kernel/bpf/cpumap.c if (!cmap->flush_list) flush_list 123 kernel/bpf/cpumap.c INIT_LIST_HEAD(per_cpu_ptr(cmap->flush_list, cpu)); flush_list 134 kernel/bpf/cpumap.c free_percpu(cmap->flush_list); flush_list 531 kernel/bpf/cpumap.c struct list_head *flush_list = per_cpu_ptr(cmap->flush_list, cpu); flush_list 533 kernel/bpf/cpumap.c while (!list_empty(flush_list)) flush_list 550 kernel/bpf/cpumap.c free_percpu(cmap->flush_list); flush_list 645 kernel/bpf/cpumap.c struct list_head *flush_list = this_cpu_ptr(rcpu->cmap->flush_list); flush_list 663 kernel/bpf/cpumap.c list_add(&bq->flush_node, flush_list); flush_list 687 kernel/bpf/cpumap.c struct list_head *flush_list = this_cpu_ptr(cmap->flush_list); flush_list 690 kernel/bpf/cpumap.c list_for_each_entry_safe(bq, tmp, flush_list, flush_node) { flush_list 78 kernel/bpf/devmap.c struct list_head __percpu *flush_list; flush_list 146 kernel/bpf/devmap.c dtab->flush_list = alloc_percpu(struct list_head); flush_list 147 kernel/bpf/devmap.c if (!dtab->flush_list) flush_list 151 kernel/bpf/devmap.c INIT_LIST_HEAD(per_cpu_ptr(dtab->flush_list, cpu)); flush_list 170 kernel/bpf/devmap.c free_percpu(dtab->flush_list); flush_list 230 kernel/bpf/devmap.c struct list_head *flush_list = per_cpu_ptr(dtab->flush_list, cpu); flush_list 232 kernel/bpf/devmap.c while (!list_empty(flush_list)) flush_list 269 kernel/bpf/devmap.c free_percpu(dtab->flush_list); flush_list 408 kernel/bpf/devmap.c struct list_head *flush_list = this_cpu_ptr(dtab->flush_list); flush_list 412 kernel/bpf/devmap.c list_for_each_entry_safe(bq, tmp, flush_list, flush_node) flush_list 440 kernel/bpf/devmap.c struct list_head *flush_list = this_cpu_ptr(obj->dtab->flush_list); flush_list 456 kernel/bpf/devmap.c list_add(&bq->flush_node, flush_list); flush_list 15 kernel/bpf/xskmap.c struct list_head __percpu *flush_list; flush_list 112 kernel/bpf/xskmap.c m->flush_list = alloc_percpu(struct list_head); flush_list 113 kernel/bpf/xskmap.c if (!m->flush_list) flush_list 117 kernel/bpf/xskmap.c INIT_LIST_HEAD(per_cpu_ptr(m->flush_list, cpu)); flush_list 127 kernel/bpf/xskmap.c free_percpu(m->flush_list); flush_list 141 kernel/bpf/xskmap.c free_percpu(m->flush_list); flush_list 179 kernel/bpf/xskmap.c struct list_head *flush_list = this_cpu_ptr(m->flush_list); flush_list 187 kernel/bpf/xskmap.c list_add(&xs->flush_node, flush_list); flush_list 195 kernel/bpf/xskmap.c struct list_head *flush_list = this_cpu_ptr(m->flush_list); flush_list 198 kernel/bpf/xskmap.c list_for_each_entry_safe(xs, tmp, flush_list, flush_node) {