gc_work           212 drivers/gpu/drm/qxl/qxl_cmd.c 		schedule_work(&qdev->gc_work);
gc_work           214 drivers/gpu/drm/qxl/qxl_cmd.c 			flush_work(&qdev->gc_work);
gc_work           271 drivers/gpu/drm/qxl/qxl_drv.h 	struct work_struct gc_work;
gc_work           104 drivers/gpu/drm/qxl/qxl_kms.c 	struct qxl_device *qdev = container_of(work, struct qxl_device, gc_work);
gc_work           274 drivers/gpu/drm/qxl/qxl_kms.c 	INIT_WORK(&qdev->gc_work, qxl_gc_work);
gc_work           302 drivers/gpu/drm/qxl/qxl_kms.c 	flush_work(&qdev->gc_work);
gc_work           217 include/net/neighbour.h 	struct delayed_work	gc_work;
gc_work            28 include/net/netfilter/nf_flow_table.h 	struct delayed_work		gc_work;
gc_work           456 net/bridge/br_device.c 	INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup);
gc_work           335 net/bridge/br_fdb.c 					     gc_work.work);
gc_work           365 net/bridge/br_fdb.c 	mod_delayed_work(system_long_wq, &br->gc_work, work_delay);
gc_work           380 net/bridge/br_if.c 	cancel_delayed_work_sync(&br->gc_work);
gc_work           148 net/bridge/br_ioctl.c 		b.gc_timer_value = br_timer_value(&br->gc_work.timer);
gc_work          1410 net/bridge/br_netlink.c 	clockval = br_timer_value(&br->gc_work.timer);
gc_work           412 net/bridge/br_private.h 	struct delayed_work		gc_work;
gc_work           601 net/bridge/br_stp.c 	mod_delayed_work(system_long_wq, &br->gc_work, 0);
gc_work            56 net/bridge/br_stp_if.c 	mod_delayed_work(system_long_wq, &br->gc_work, HZ / 10);
gc_work            87 net/bridge/br_stp_if.c 	cancel_delayed_work_sync(&br->gc_work);
gc_work           263 net/bridge/br_sysfs_br.c 	return sprintf(buf, "%ld\n", br_timer_value(&br->gc_work.timer));
gc_work           887 net/core/neighbour.c 	struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work);
gc_work           962 net/core/neighbour.c 	queue_delayed_work(system_power_efficient_wq, &tbl->gc_work,
gc_work          1711 net/core/neighbour.c 	INIT_DEFERRABLE_WORK(&tbl->gc_work, neigh_periodic_work);
gc_work          1712 net/core/neighbour.c 	queue_delayed_work(system_power_efficient_wq, &tbl->gc_work,
gc_work          1729 net/core/neighbour.c 	cancel_delayed_work_sync(&tbl->gc_work);
gc_work            62 net/netfilter/nf_conncount.c 	struct work_struct gc_work;
gc_work           296 net/netfilter/nf_conncount.c 	schedule_work(&data->gc_work);
gc_work           446 net/netfilter/nf_conncount.c 	struct nf_conncount_data *data = container_of(work, struct nf_conncount_data, gc_work);
gc_work           544 net/netfilter/nf_conncount.c 	INIT_WORK(&data->gc_work, tree_gc_worker);
gc_work           580 net/netfilter/nf_conncount.c 	cancel_work_sync(&data->gc_work);
gc_work          1226 net/netfilter/nf_conntrack_core.c 	struct conntrack_gc_work *gc_work;
gc_work          1230 net/netfilter/nf_conntrack_core.c 	gc_work = container_of(work, struct conntrack_gc_work, dwork.work);
gc_work          1233 net/netfilter/nf_conntrack_core.c 	i = gc_work->last_bucket;
gc_work          1234 net/netfilter/nf_conntrack_core.c 	if (gc_work->early_drop)
gc_work          1298 net/netfilter/nf_conntrack_core.c 	if (gc_work->exiting)
gc_work          1320 net/netfilter/nf_conntrack_core.c 		gc_work->next_gc_run = min_interval;
gc_work          1326 net/netfilter/nf_conntrack_core.c 		gc_work->next_gc_run += min_interval;
gc_work          1327 net/netfilter/nf_conntrack_core.c 		if (gc_work->next_gc_run > max)
gc_work          1328 net/netfilter/nf_conntrack_core.c 			gc_work->next_gc_run = max;
gc_work          1331 net/netfilter/nf_conntrack_core.c 	next_run = gc_work->next_gc_run;
gc_work          1332 net/netfilter/nf_conntrack_core.c 	gc_work->last_bucket = i;
gc_work          1333 net/netfilter/nf_conntrack_core.c 	gc_work->early_drop = false;
gc_work          1334 net/netfilter/nf_conntrack_core.c 	queue_delayed_work(system_power_efficient_wq, &gc_work->dwork, next_run);
gc_work          1337 net/netfilter/nf_conntrack_core.c static void conntrack_gc_work_init(struct conntrack_gc_work *gc_work)
gc_work          1339 net/netfilter/nf_conntrack_core.c 	INIT_DEFERRABLE_WORK(&gc_work->dwork, gc_worker);
gc_work          1340 net/netfilter/nf_conntrack_core.c 	gc_work->next_gc_run = HZ;
gc_work          1341 net/netfilter/nf_conntrack_core.c 	gc_work->exiting = false;
gc_work           342 net/netfilter/nf_flow_table_core.c 	flow_table = container_of(work, struct nf_flowtable, gc_work.work);
gc_work           344 net/netfilter/nf_flow_table_core.c 	queue_delayed_work(system_power_efficient_wq, &flow_table->gc_work, HZ);
gc_work           467 net/netfilter/nf_flow_table_core.c 	INIT_DEFERRABLE_WORK(&flowtable->gc_work, nf_flow_offload_work_gc);
gc_work           475 net/netfilter/nf_flow_table_core.c 			   &flowtable->gc_work, HZ);
gc_work           506 net/netfilter/nf_flow_table_core.c 	flush_delayed_work(&flowtable->gc_work);
gc_work           525 net/netfilter/nf_flow_table_core.c 	cancel_delayed_work_sync(&flow_table->gc_work);
gc_work            26 net/netfilter/nft_set_hash.c 	struct delayed_work		gc_work;
gc_work           304 net/netfilter/nft_set_hash.c 	priv = container_of(work, struct nft_rhash, gc_work.work);
gc_work           341 net/netfilter/nft_set_hash.c 	queue_delayed_work(system_power_efficient_wq, &priv->gc_work,
gc_work           355 net/netfilter/nft_set_hash.c 	queue_delayed_work(system_power_efficient_wq, &priv->gc_work,
gc_work           374 net/netfilter/nft_set_hash.c 	INIT_DEFERRABLE_WORK(&priv->gc_work, nft_rhash_gc);
gc_work           390 net/netfilter/nft_set_hash.c 	cancel_delayed_work_sync(&priv->gc_work);
gc_work            22 net/netfilter/nft_set_rbtree.c 	struct delayed_work	gc_work;
gc_work           393 net/netfilter/nft_set_rbtree.c 	priv = container_of(work, struct nft_rbtree, gc_work.work);
gc_work           439 net/netfilter/nft_set_rbtree.c 	queue_delayed_work(system_power_efficient_wq, &priv->gc_work,
gc_work           459 net/netfilter/nft_set_rbtree.c 	INIT_DEFERRABLE_WORK(&priv->gc_work, nft_rbtree_gc);
gc_work           461 net/netfilter/nft_set_rbtree.c 		queue_delayed_work(system_power_efficient_wq, &priv->gc_work,
gc_work           473 net/netfilter/nft_set_rbtree.c 	cancel_delayed_work_sync(&priv->gc_work);
gc_work           128 net/netfilter/xt_hashlimit.c 	struct delayed_work gc_work;
gc_work           352 net/netfilter/xt_hashlimit.c 	INIT_DEFERRABLE_WORK(&hinfo->gc_work, htable_gc);
gc_work           353 net/netfilter/xt_hashlimit.c 	queue_delayed_work(system_power_efficient_wq, &hinfo->gc_work,
gc_work           383 net/netfilter/xt_hashlimit.c 	ht = container_of(work, struct xt_hashlimit_htable, gc_work.work);
gc_work           388 net/netfilter/xt_hashlimit.c 			   &ht->gc_work, msecs_to_jiffies(ht->cfg.gc_interval));
gc_work           429 net/netfilter/xt_hashlimit.c 		cancel_delayed_work_sync(&hinfo->gc_work);