gc_work 212 drivers/gpu/drm/qxl/qxl_cmd.c schedule_work(&qdev->gc_work); gc_work 214 drivers/gpu/drm/qxl/qxl_cmd.c flush_work(&qdev->gc_work); gc_work 271 drivers/gpu/drm/qxl/qxl_drv.h struct work_struct gc_work; gc_work 104 drivers/gpu/drm/qxl/qxl_kms.c struct qxl_device *qdev = container_of(work, struct qxl_device, gc_work); gc_work 274 drivers/gpu/drm/qxl/qxl_kms.c INIT_WORK(&qdev->gc_work, qxl_gc_work); gc_work 302 drivers/gpu/drm/qxl/qxl_kms.c flush_work(&qdev->gc_work); gc_work 217 include/net/neighbour.h struct delayed_work gc_work; gc_work 28 include/net/netfilter/nf_flow_table.h struct delayed_work gc_work; gc_work 456 net/bridge/br_device.c INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup); gc_work 335 net/bridge/br_fdb.c gc_work.work); gc_work 365 net/bridge/br_fdb.c mod_delayed_work(system_long_wq, &br->gc_work, work_delay); gc_work 380 net/bridge/br_if.c cancel_delayed_work_sync(&br->gc_work); gc_work 148 net/bridge/br_ioctl.c b.gc_timer_value = br_timer_value(&br->gc_work.timer); gc_work 1410 net/bridge/br_netlink.c clockval = br_timer_value(&br->gc_work.timer); gc_work 412 net/bridge/br_private.h struct delayed_work gc_work; gc_work 601 net/bridge/br_stp.c mod_delayed_work(system_long_wq, &br->gc_work, 0); gc_work 56 net/bridge/br_stp_if.c mod_delayed_work(system_long_wq, &br->gc_work, HZ / 10); gc_work 87 net/bridge/br_stp_if.c cancel_delayed_work_sync(&br->gc_work); gc_work 263 net/bridge/br_sysfs_br.c return sprintf(buf, "%ld\n", br_timer_value(&br->gc_work.timer)); gc_work 887 net/core/neighbour.c struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); gc_work 962 net/core/neighbour.c queue_delayed_work(system_power_efficient_wq, &tbl->gc_work, gc_work 1711 net/core/neighbour.c INIT_DEFERRABLE_WORK(&tbl->gc_work, neigh_periodic_work); gc_work 1712 net/core/neighbour.c queue_delayed_work(system_power_efficient_wq, &tbl->gc_work, gc_work 1729 net/core/neighbour.c cancel_delayed_work_sync(&tbl->gc_work); gc_work 62 net/netfilter/nf_conncount.c struct work_struct gc_work; gc_work 296 net/netfilter/nf_conncount.c schedule_work(&data->gc_work); gc_work 446 net/netfilter/nf_conncount.c struct nf_conncount_data *data = container_of(work, struct nf_conncount_data, gc_work); gc_work 544 net/netfilter/nf_conncount.c INIT_WORK(&data->gc_work, tree_gc_worker); gc_work 580 net/netfilter/nf_conncount.c cancel_work_sync(&data->gc_work); gc_work 1226 net/netfilter/nf_conntrack_core.c struct conntrack_gc_work *gc_work; gc_work 1230 net/netfilter/nf_conntrack_core.c gc_work = container_of(work, struct conntrack_gc_work, dwork.work); gc_work 1233 net/netfilter/nf_conntrack_core.c i = gc_work->last_bucket; gc_work 1234 net/netfilter/nf_conntrack_core.c if (gc_work->early_drop) gc_work 1298 net/netfilter/nf_conntrack_core.c if (gc_work->exiting) gc_work 1320 net/netfilter/nf_conntrack_core.c gc_work->next_gc_run = min_interval; gc_work 1326 net/netfilter/nf_conntrack_core.c gc_work->next_gc_run += min_interval; gc_work 1327 net/netfilter/nf_conntrack_core.c if (gc_work->next_gc_run > max) gc_work 1328 net/netfilter/nf_conntrack_core.c gc_work->next_gc_run = max; gc_work 1331 net/netfilter/nf_conntrack_core.c next_run = gc_work->next_gc_run; gc_work 1332 net/netfilter/nf_conntrack_core.c gc_work->last_bucket = i; gc_work 1333 net/netfilter/nf_conntrack_core.c gc_work->early_drop = false; gc_work 1334 net/netfilter/nf_conntrack_core.c queue_delayed_work(system_power_efficient_wq, &gc_work->dwork, next_run); gc_work 1337 net/netfilter/nf_conntrack_core.c static void conntrack_gc_work_init(struct conntrack_gc_work *gc_work) gc_work 1339 net/netfilter/nf_conntrack_core.c INIT_DEFERRABLE_WORK(&gc_work->dwork, gc_worker); gc_work 1340 net/netfilter/nf_conntrack_core.c gc_work->next_gc_run = HZ; gc_work 1341 net/netfilter/nf_conntrack_core.c gc_work->exiting = false; gc_work 342 net/netfilter/nf_flow_table_core.c flow_table = container_of(work, struct nf_flowtable, gc_work.work); gc_work 344 net/netfilter/nf_flow_table_core.c queue_delayed_work(system_power_efficient_wq, &flow_table->gc_work, HZ); gc_work 467 net/netfilter/nf_flow_table_core.c INIT_DEFERRABLE_WORK(&flowtable->gc_work, nf_flow_offload_work_gc); gc_work 475 net/netfilter/nf_flow_table_core.c &flowtable->gc_work, HZ); gc_work 506 net/netfilter/nf_flow_table_core.c flush_delayed_work(&flowtable->gc_work); gc_work 525 net/netfilter/nf_flow_table_core.c cancel_delayed_work_sync(&flow_table->gc_work); gc_work 26 net/netfilter/nft_set_hash.c struct delayed_work gc_work; gc_work 304 net/netfilter/nft_set_hash.c priv = container_of(work, struct nft_rhash, gc_work.work); gc_work 341 net/netfilter/nft_set_hash.c queue_delayed_work(system_power_efficient_wq, &priv->gc_work, gc_work 355 net/netfilter/nft_set_hash.c queue_delayed_work(system_power_efficient_wq, &priv->gc_work, gc_work 374 net/netfilter/nft_set_hash.c INIT_DEFERRABLE_WORK(&priv->gc_work, nft_rhash_gc); gc_work 390 net/netfilter/nft_set_hash.c cancel_delayed_work_sync(&priv->gc_work); gc_work 22 net/netfilter/nft_set_rbtree.c struct delayed_work gc_work; gc_work 393 net/netfilter/nft_set_rbtree.c priv = container_of(work, struct nft_rbtree, gc_work.work); gc_work 439 net/netfilter/nft_set_rbtree.c queue_delayed_work(system_power_efficient_wq, &priv->gc_work, gc_work 459 net/netfilter/nft_set_rbtree.c INIT_DEFERRABLE_WORK(&priv->gc_work, nft_rbtree_gc); gc_work 461 net/netfilter/nft_set_rbtree.c queue_delayed_work(system_power_efficient_wq, &priv->gc_work, gc_work 473 net/netfilter/nft_set_rbtree.c cancel_delayed_work_sync(&priv->gc_work); gc_work 128 net/netfilter/xt_hashlimit.c struct delayed_work gc_work; gc_work 352 net/netfilter/xt_hashlimit.c INIT_DEFERRABLE_WORK(&hinfo->gc_work, htable_gc); gc_work 353 net/netfilter/xt_hashlimit.c queue_delayed_work(system_power_efficient_wq, &hinfo->gc_work, gc_work 383 net/netfilter/xt_hashlimit.c ht = container_of(work, struct xt_hashlimit_htable, gc_work.work); gc_work 388 net/netfilter/xt_hashlimit.c &ht->gc_work, msecs_to_jiffies(ht->cfg.gc_interval)); gc_work 429 net/netfilter/xt_hashlimit.c cancel_delayed_work_sync(&hinfo->gc_work);