/linux/drivers/net/wireguard/ |
A D | ratelimiter.c | 20 static DECLARE_DEFERRABLE_WORK(gc_work, wg_ratelimiter_gc_entries); 81 queue_delayed_work(system_power_efficient_wq, &gc_work, HZ); in wg_ratelimiter_gc_entries() 191 queue_delayed_work(system_power_efficient_wq, &gc_work, HZ); in wg_ratelimiter_init() 211 cancel_delayed_work_sync(&gc_work); in wg_ratelimiter_uninit()
|
/linux/drivers/gpu/drm/qxl/ |
A D | qxl_kms.c | 104 struct qxl_device *qdev = container_of(work, struct qxl_device, gc_work); in qxl_gc_work() 264 INIT_WORK(&qdev->gc_work, qxl_gc_work); in qxl_device_init() 292 if (!qdev->gc_work.func) in qxl_device_fini() 312 flush_work(&qdev->gc_work); in qxl_device_fini()
|
A D | qxl_cmd.c | 212 schedule_work(&qdev->gc_work); in qxl_queue_garbage_collect() 214 flush_work(&qdev->gc_work); in qxl_queue_garbage_collect()
|
A D | qxl_drv.h | 253 struct work_struct gc_work; member
|
/linux/net/netfilter/ |
A D | nft_set_rbtree.c | 22 struct delayed_work gc_work; member 501 priv = container_of(work, struct nft_rbtree, gc_work.work); in nft_rbtree_gc() 553 queue_delayed_work(system_power_efficient_wq, &priv->gc_work, in nft_rbtree_gc() 573 INIT_DEFERRABLE_WORK(&priv->gc_work, nft_rbtree_gc); in nft_rbtree_init() 575 queue_delayed_work(system_power_efficient_wq, &priv->gc_work, in nft_rbtree_init() 587 cancel_delayed_work_sync(&priv->gc_work); in nft_rbtree_destroy()
|
A D | nf_conncount.c | 62 struct work_struct gc_work; member 296 schedule_work(&data->gc_work); in schedule_gc_worker() 446 struct nf_conncount_data *data = container_of(work, struct nf_conncount_data, gc_work); in tree_gc_worker() 544 INIT_WORK(&data->gc_work, tree_gc_worker); in nf_conncount_init() 580 cancel_work_sync(&data->gc_work); in nf_conncount_destroy()
|
A D | nf_flow_table_core.c | 481 flow_table = container_of(work, struct nf_flowtable, gc_work.work); in nf_flow_offload_work_gc() 483 queue_delayed_work(system_power_efficient_wq, &flow_table->gc_work, HZ); in nf_flow_offload_work_gc() 578 INIT_DELAYED_WORK(&flowtable->gc_work, nf_flow_offload_work_gc); in nf_flow_table_init() 588 &flowtable->gc_work, HZ); in nf_flow_table_init() 617 flush_delayed_work(&flowtable->gc_work); in nf_flow_table_gc_cleanup() 638 cancel_delayed_work_sync(&flow_table->gc_work); in nf_flow_table_free()
|
A D | nft_set_hash.c | 26 struct delayed_work gc_work; member 321 priv = container_of(work, struct nft_rhash, gc_work.work); in nft_rhash_gc() 361 queue_delayed_work(system_power_efficient_wq, &priv->gc_work, in nft_rhash_gc() 375 queue_delayed_work(system_power_efficient_wq, &priv->gc_work, in nft_rhash_gc_init() 394 INIT_DEFERRABLE_WORK(&priv->gc_work, nft_rhash_gc); in nft_rhash_init() 410 cancel_delayed_work_sync(&priv->gc_work); in nft_rhash_destroy()
|
A D | nf_conntrack_core.c | 1426 struct conntrack_gc_work *gc_work; in gc_worker() local 1427 gc_work = container_of(work, struct conntrack_gc_work, dwork.work); in gc_worker() 1429 i = gc_work->next_bucket; in gc_worker() 1430 if (gc_work->early_drop) in gc_worker() 1495 gc_work->next_bucket = i; in gc_worker() 1501 if (gc_work->exiting) in gc_worker() 1512 gc_work->early_drop = false; in gc_worker() 1513 gc_work->next_bucket = 0; in gc_worker() 1518 static void conntrack_gc_work_init(struct conntrack_gc_work *gc_work) in conntrack_gc_work_init() argument 1520 INIT_DEFERRABLE_WORK(&gc_work->dwork, gc_worker); in conntrack_gc_work_init() [all …]
|
A D | xt_hashlimit.c | 128 struct delayed_work gc_work; member 352 INIT_DEFERRABLE_WORK(&hinfo->gc_work, htable_gc); in htable_create() 353 queue_delayed_work(system_power_efficient_wq, &hinfo->gc_work, in htable_create() 383 ht = container_of(work, struct xt_hashlimit_htable, gc_work.work); in htable_gc() 388 &ht->gc_work, msecs_to_jiffies(ht->cfg.gc_interval)); in htable_gc() 429 cancel_delayed_work_sync(&hinfo->gc_work); in htable_put()
|
/linux/net/bridge/ |
A D | br_stp_if.c | 56 mod_delayed_work(system_long_wq, &br->gc_work, HZ / 10); in br_stp_enable_bridge() 87 cancel_delayed_work_sync(&br->gc_work); in br_stp_disable_bridge()
|
A D | br_ioctl.c | 166 b.gc_timer_value = br_timer_value(&br->gc_work.timer); in br_dev_siocdevprivate()
|
A D | br_device.c | 532 INIT_DELAYED_WORK(&br->gc_work, br_fdb_cleanup); in br_dev_setup()
|
A D | br_fdb.c | 518 gc_work.work); in br_fdb_cleanup() 558 mod_delayed_work(system_long_wq, &br->gc_work, work_delay); in br_fdb_cleanup()
|
A D | br_if.c | 383 cancel_delayed_work_sync(&br->gc_work); in br_dev_delete()
|
A D | br_stp.c | 637 mod_delayed_work(system_long_wq, &br->gc_work, 0); in br_set_ageing_time()
|
A D | br_sysfs_br.c | 295 return sprintf(buf, "%ld\n", br_timer_value(&br->gc_work.timer)); in gc_timer_show()
|
A D | br_private.h | 518 struct delayed_work gc_work; member
|
A D | br_netlink.c | 1539 clockval = br_timer_value(&br->gc_work.timer); in br_fill_info()
|
/linux/include/net/netfilter/ |
A D | nf_flow_table.h | 76 struct delayed_work gc_work; member
|
/linux/include/net/ |
A D | neighbour.h | 219 struct delayed_work gc_work; member
|
/linux/net/core/ |
A D | neighbour.c | 918 struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); in neigh_periodic_work() 993 queue_delayed_work(system_power_efficient_wq, &tbl->gc_work, in neigh_periodic_work() 1760 INIT_DEFERRABLE_WORK(&tbl->gc_work, neigh_periodic_work); in neigh_table_init() 1761 queue_delayed_work(system_power_efficient_wq, &tbl->gc_work, in neigh_table_init() 1782 cancel_delayed_work_sync(&tbl->gc_work); in neigh_table_clear()
|