flush_list        466 drivers/md/dm-log-userspace-base.c static int flush_one_by_one(struct log_c *lc, struct list_head *flush_list)
flush_list        471 drivers/md/dm-log-userspace-base.c 	list_for_each_entry(fe, flush_list, list) {
flush_list        483 drivers/md/dm-log-userspace-base.c static int flush_by_group(struct log_c *lc, struct list_head *flush_list,
flush_list        496 drivers/md/dm-log-userspace-base.c 	while (!list_empty(flush_list)) {
flush_list        499 drivers/md/dm-log-userspace-base.c 		list_for_each_entry_safe(fe, tmp_fe, flush_list, list) {
flush_list        529 drivers/md/dm-log-userspace-base.c 				list_splice_init(&tmp_list, flush_list);
flush_list        530 drivers/md/dm-log-userspace-base.c 				r = flush_one_by_one(lc, flush_list);
flush_list        540 drivers/md/dm-log-userspace-base.c 	list_splice_init(&tmp_list, flush_list);
flush_list        176 drivers/md/dm-writecache.c 	struct bio_list flush_list;
flush_list       1110 drivers/md/dm-writecache.c 		bio = bio_list_pop(&wc->flush_list);
flush_list       1144 drivers/md/dm-writecache.c 	if (bio_list_empty(&wc->flush_list))
flush_list       1146 drivers/md/dm-writecache.c 	bio_list_add(&wc->flush_list, bio);
flush_list       2114 drivers/md/dm-writecache.c 		bio_list_init(&wc->flush_list);
flush_list         64 drivers/md/dm-zoned-target.c 	struct bio_list		flush_list;
flush_list        510 drivers/md/dm-zoned-target.c 		bio = bio_list_pop(&dmz->flush_list);
flush_list        655 drivers/md/dm-zoned-target.c 		bio_list_add(&dmz->flush_list, bio);
flush_list        824 drivers/md/dm-zoned-target.c 	bio_list_init(&dmz->flush_list);
flush_list         75 kernel/bpf/cpumap.c 	struct list_head __percpu *flush_list;
flush_list        118 kernel/bpf/cpumap.c 	cmap->flush_list = alloc_percpu(struct list_head);
flush_list        119 kernel/bpf/cpumap.c 	if (!cmap->flush_list)
flush_list        123 kernel/bpf/cpumap.c 		INIT_LIST_HEAD(per_cpu_ptr(cmap->flush_list, cpu));
flush_list        134 kernel/bpf/cpumap.c 	free_percpu(cmap->flush_list);
flush_list        531 kernel/bpf/cpumap.c 		struct list_head *flush_list = per_cpu_ptr(cmap->flush_list, cpu);
flush_list        533 kernel/bpf/cpumap.c 		while (!list_empty(flush_list))
flush_list        550 kernel/bpf/cpumap.c 	free_percpu(cmap->flush_list);
flush_list        645 kernel/bpf/cpumap.c 	struct list_head *flush_list = this_cpu_ptr(rcpu->cmap->flush_list);
flush_list        663 kernel/bpf/cpumap.c 		list_add(&bq->flush_node, flush_list);
flush_list        687 kernel/bpf/cpumap.c 	struct list_head *flush_list = this_cpu_ptr(cmap->flush_list);
flush_list        690 kernel/bpf/cpumap.c 	list_for_each_entry_safe(bq, tmp, flush_list, flush_node) {
flush_list         78 kernel/bpf/devmap.c 	struct list_head __percpu *flush_list;
flush_list        146 kernel/bpf/devmap.c 	dtab->flush_list = alloc_percpu(struct list_head);
flush_list        147 kernel/bpf/devmap.c 	if (!dtab->flush_list)
flush_list        151 kernel/bpf/devmap.c 		INIT_LIST_HEAD(per_cpu_ptr(dtab->flush_list, cpu));
flush_list        170 kernel/bpf/devmap.c 	free_percpu(dtab->flush_list);
flush_list        230 kernel/bpf/devmap.c 		struct list_head *flush_list = per_cpu_ptr(dtab->flush_list, cpu);
flush_list        232 kernel/bpf/devmap.c 		while (!list_empty(flush_list))
flush_list        269 kernel/bpf/devmap.c 	free_percpu(dtab->flush_list);
flush_list        408 kernel/bpf/devmap.c 	struct list_head *flush_list = this_cpu_ptr(dtab->flush_list);
flush_list        412 kernel/bpf/devmap.c 	list_for_each_entry_safe(bq, tmp, flush_list, flush_node)
flush_list        440 kernel/bpf/devmap.c 	struct list_head *flush_list = this_cpu_ptr(obj->dtab->flush_list);
flush_list        456 kernel/bpf/devmap.c 		list_add(&bq->flush_node, flush_list);
flush_list         15 kernel/bpf/xskmap.c 	struct list_head __percpu *flush_list;
flush_list        112 kernel/bpf/xskmap.c 	m->flush_list = alloc_percpu(struct list_head);
flush_list        113 kernel/bpf/xskmap.c 	if (!m->flush_list)
flush_list        117 kernel/bpf/xskmap.c 		INIT_LIST_HEAD(per_cpu_ptr(m->flush_list, cpu));
flush_list        127 kernel/bpf/xskmap.c 	free_percpu(m->flush_list);
flush_list        141 kernel/bpf/xskmap.c 	free_percpu(m->flush_list);
flush_list        179 kernel/bpf/xskmap.c 	struct list_head *flush_list = this_cpu_ptr(m->flush_list);
flush_list        187 kernel/bpf/xskmap.c 		list_add(&xs->flush_node, flush_list);
flush_list        195 kernel/bpf/xskmap.c 	struct list_head *flush_list = this_cpu_ptr(m->flush_list);
flush_list        198 kernel/bpf/xskmap.c 	list_for_each_entry_safe(xs, tmp, flush_list, flush_node) {