to_free 228 arch/um/kernel/irq.c struct irq_entry *to_free; to_free 246 arch/um/kernel/irq.c to_free = walk; to_free 248 arch/um/kernel/irq.c to_free = NULL; to_free 251 arch/um/kernel/irq.c kfree(to_free); to_free 290 arch/um/kernel/irq.c struct irq_fd *to_free; to_free 304 arch/um/kernel/irq.c to_free = irq_entry->irq_array[i]; to_free 307 arch/um/kernel/irq.c if (to_free->active) to_free 308 arch/um/kernel/irq.c to_free->purge = true; to_free 310 arch/um/kernel/irq.c kfree(to_free); to_free 318 arch/um/kernel/irq.c struct irq_entry *to_free; to_free 322 arch/um/kernel/irq.c to_free = get_irq_entry_by_fd(fd); to_free 323 arch/um/kernel/irq.c if (to_free != NULL) { to_free 325 arch/um/kernel/irq.c to_free, to_free 338 arch/um/kernel/irq.c struct irq_entry *to_free; to_free 342 arch/um/kernel/irq.c to_free = active_fds; to_free 343 arch/um/kernel/irq.c while (to_free != NULL) { to_free 345 arch/um/kernel/irq.c to_free, to_free 350 arch/um/kernel/irq.c to_free = to_free->next; to_free 359 arch/um/kernel/irq.c struct irq_entry *to_free; to_free 364 arch/um/kernel/irq.c to_free = get_irq_entry_by_fd(fd); to_free 365 arch/um/kernel/irq.c if (to_free != NULL) { to_free 367 arch/um/kernel/irq.c to_free, to_free 387 arch/um/kernel/irq.c struct irq_entry *to_free; to_free 394 arch/um/kernel/irq.c to_free = active_fds; to_free 395 arch/um/kernel/irq.c while (to_free != NULL) { to_free 397 arch/um/kernel/irq.c to_free, to_free 402 arch/um/kernel/irq.c to_free = to_free->next; to_free 550 drivers/android/binder_alloc.c bool to_free = true; to_free 555 drivers/android/binder_alloc.c to_free = false; to_free 565 drivers/android/binder_alloc.c to_free = false; to_free 578 drivers/android/binder_alloc.c to_free = false; to_free 581 drivers/android/binder_alloc.c if (to_free) { to_free 75 drivers/infiniband/sw/siw/siw_mem.c int to_free = min_t(int, PAGES_PER_CHUNK, num_pages); to_free 77 drivers/infiniband/sw/siw/siw_mem.c siw_free_plist(&umem->page_chunk[i], to_free, to_free 80 drivers/infiniband/sw/siw/siw_mem.c num_pages -= to_free; to_free 116 drivers/media/pci/ivtv/ivtv-queue.c int to_free = to == &s->q_free; to_free 166 drivers/media/pci/ivtv/ivtv-queue.c ivtv_queue_move_buf(s, from, to, to_free); to_free 1245 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c u32 occup, to_free, freed, freed_start; to_free 1248 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c occup = to_free = REG_RD(bp, regs->lines_occup); to_free 1254 drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c while (occup && ((u32)SUB_S32(freed, freed_start) < to_free)) { to_free 257 drivers/scsi/cxlflash/vlun.c static int ba_free(struct ba_lun *ba_lun, u64 to_free) to_free 264 drivers/scsi/cxlflash/vlun.c if (validate_alloc(bali, to_free)) { to_free 266 drivers/scsi/cxlflash/vlun.c __func__, to_free, ba_lun->lun_id); to_free 271 drivers/scsi/cxlflash/vlun.c "free_aun_cnt=%llx\n", __func__, to_free, ba_lun->lun_id, to_free 274 drivers/scsi/cxlflash/vlun.c if (bali->aun_clone_map[to_free] > 0) { to_free 276 drivers/scsi/cxlflash/vlun.c __func__, to_free, ba_lun->lun_id, to_free 277 drivers/scsi/cxlflash/vlun.c bali->aun_clone_map[to_free]); to_free 278 drivers/scsi/cxlflash/vlun.c bali->aun_clone_map[to_free]--; to_free 282 drivers/scsi/cxlflash/vlun.c idx = to_free / BITS_PER_LONG; to_free 283 drivers/scsi/cxlflash/vlun.c bit_pos = to_free % BITS_PER_LONG; to_free 304 drivers/tty/serial/8250/8250_hp300.c struct hp300_port *port, *to_free; to_free 308 drivers/tty/serial/8250/8250_hp300.c to_free = port; to_free 310 drivers/tty/serial/8250/8250_hp300.c kfree(to_free); to_free 218 drivers/usb/gadget/function/f_fs.c const void *to_free; to_free 844 drivers/usb/gadget/function/f_fs.c kfree(io_data->to_free); to_free 1240 drivers/usb/gadget/function/f_fs.c p->to_free = dup_iter(&p->data, to, GFP_KERNEL); to_free 1241 drivers/usb/gadget/function/f_fs.c if (!p->to_free) { to_free 1247 drivers/usb/gadget/function/f_fs.c p->to_free = NULL; to_free 1261 drivers/usb/gadget/function/f_fs.c kfree(p->to_free); to_free 435 drivers/usb/gadget/legacy/inode.c const void *to_free; to_free 475 drivers/usb/gadget/legacy/inode.c kfree(priv->to_free); to_free 494 drivers/usb/gadget/legacy/inode.c if (priv->to_free == NULL || unlikely(req->actual == 0)) { to_free 496 drivers/usb/gadget/legacy/inode.c kfree(priv->to_free); to_free 566 drivers/usb/gadget/legacy/inode.c kfree(priv->to_free); to_free 614 drivers/usb/gadget/legacy/inode.c priv->to_free = dup_iter(&priv->to, to, GFP_KERNEL); to_free 615 drivers/usb/gadget/legacy/inode.c if (!priv->to_free) { to_free 131 fs/btrfs/delayed-ref.c u64 to_free = 0; to_free 143 fs/btrfs/delayed-ref.c to_free = num_bytes - delta; to_free 147 fs/btrfs/delayed-ref.c to_free = num_bytes; to_free 160 fs/btrfs/delayed-ref.c if (to_free) to_free 162 fs/btrfs/delayed-ref.c delayed_refs_rsv->space_info, to_free); to_free 2435 fs/btrfs/free-space-cache.c u64 to_free = min(bytes, info->bytes); to_free 2437 fs/btrfs/free-space-cache.c info->bytes -= to_free; to_free 2438 fs/btrfs/free-space-cache.c info->offset += to_free; to_free 2446 fs/btrfs/free-space-cache.c offset += to_free; to_free 2447 fs/btrfs/free-space-cache.c bytes -= to_free; to_free 1110 fs/ceph/snap.c LIST_HEAD(to_free); to_free 1122 fs/ceph/snap.c list_move(&sm->lru, &to_free); to_free 1126 fs/ceph/snap.c while (!list_empty(&to_free)) { to_free 1127 fs/ceph/snap.c sm = list_first_entry(&to_free, struct ceph_snapid_map, lru); to_free 1139 fs/ceph/snap.c LIST_HEAD(to_free); to_free 1146 fs/ceph/snap.c list_move(&sm->lru, &to_free); to_free 1150 fs/ceph/snap.c while (!list_empty(&to_free)) { to_free 1151 fs/ceph/snap.c sm = list_first_entry(&to_free, struct ceph_snapid_map, lru); to_free 2642 fs/ext4/ext4.h extern void ext4_da_release_space(struct inode *inode, int to_free); to_free 2689 fs/ext4/ext4.h extern void ext4_kvfree_array_rcu(void *to_free); to_free 1633 fs/ext4/inode.c void ext4_da_release_space(struct inode *inode, int to_free) to_free 1638 fs/ext4/inode.c if (!to_free) to_free 1643 fs/ext4/inode.c trace_ext4_da_release_space(inode, to_free); to_free 1644 fs/ext4/inode.c if (unlikely(to_free > ei->i_reserved_data_blocks)) { to_free 1653 fs/ext4/inode.c "data blocks", inode->i_ino, to_free, to_free 1656 fs/ext4/inode.c to_free = ei->i_reserved_data_blocks; to_free 1658 fs/ext4/inode.c ei->i_reserved_data_blocks -= to_free; to_free 1661 fs/ext4/inode.c percpu_counter_sub(&sbi->s_dirtyclusters_counter, to_free); to_free 1665 fs/ext4/inode.c dquot_release_reservation_block(inode, EXT4_C2B(sbi, to_free)); to_free 34 fs/ext4/resize.c void ext4_kvfree_array_rcu(void *to_free) to_free 39 fs/ext4/resize.c ptr->ptr = to_free; to_free 44 fs/ext4/resize.c kvfree(to_free); to_free 5309 fs/ext4/super.c char *to_free[EXT4_MAXQUOTAS]; to_free 5560 fs/ext4/super.c to_free[i] = get_qf_name(sb, sbi, i); to_free 5565 fs/ext4/super.c kfree(to_free[i]); to_free 770 fs/io_uring.c int to_free; to_free 772 fs/io_uring.c to_free = 0; to_free 788 fs/io_uring.c reqs[to_free++] = req; to_free 789 fs/io_uring.c if (to_free == ARRAY_SIZE(reqs)) to_free 790 fs/io_uring.c io_free_req_many(ctx, reqs, &to_free); to_free 798 fs/io_uring.c io_free_req_many(ctx, reqs, &to_free); to_free 235 fs/nfsd/nfs4xdr.c tb->next = argp->to_free; to_free 236 fs/nfsd/nfs4xdr.c argp->to_free = tb; to_free 4543 fs/nfsd/nfs4xdr.c while (args->to_free) { to_free 4544 fs/nfsd/nfs4xdr.c struct svcxdr_tmpbuf *tb = args->to_free; to_free 4545 fs/nfsd/nfs4xdr.c args->to_free = tb->next; to_free 4567 fs/nfsd/nfs4xdr.c args->to_free = NULL; to_free 654 fs/nfsd/xdr4.h struct svcxdr_tmpbuf *to_free; to_free 724 fs/notify/mark.c LIST_HEAD(to_free); to_free 725 fs/notify/mark.c struct list_head *head = &to_free; to_free 744 fs/notify/mark.c list_move(&mark->g_list, &to_free); to_free 838 fs/ntfs/lcnalloc.c s64 delta, to_free, total_freed, real_freed; to_free 886 fs/ntfs/lcnalloc.c to_free = rl->length - delta; to_free 887 fs/ntfs/lcnalloc.c if (count >= 0 && to_free > count) to_free 888 fs/ntfs/lcnalloc.c to_free = count; to_free 893 fs/ntfs/lcnalloc.c to_free, likely(!is_rollback) ? 0 : 1); to_free 901 fs/ntfs/lcnalloc.c real_freed = to_free; to_free 906 fs/ntfs/lcnalloc.c count -= to_free; to_free 909 fs/ntfs/lcnalloc.c total_freed = to_free; to_free 943 fs/ntfs/lcnalloc.c to_free = rl->length; to_free 944 fs/ntfs/lcnalloc.c if (count >= 0 && to_free > count) to_free 945 fs/ntfs/lcnalloc.c to_free = count; to_free 950 fs/ntfs/lcnalloc.c to_free, likely(!is_rollback) ? 0 : 1); to_free 958 fs/ntfs/lcnalloc.c real_freed += to_free; to_free 962 fs/ntfs/lcnalloc.c count -= to_free; to_free 965 fs/ntfs/lcnalloc.c total_freed += to_free; to_free 876 fs/ufs/inode.c static inline void free_data(struct to_free *ctx, u64 from, unsigned count) to_free 895 fs/ufs/inode.c struct to_free ctx = {.inode = inode}; to_free 994 fs/ufs/inode.c struct to_free ctx = {.inode = inode}; to_free 1028 fs/ufs/inode.c struct to_free ctx = {.inode = inode}; to_free 60 include/net/sch_generic.h struct sk_buff **to_free); to_free 244 include/net/sch_generic.h struct sk_buff **to_free); to_free 793 include/net/sch_generic.h struct sk_buff **to_free) to_free 796 include/net/sch_generic.h return sch->enqueue(skb, sch, to_free); to_free 1017 include/net/sch_generic.h static inline void __qdisc_drop(struct sk_buff *skb, struct sk_buff **to_free) to_free 1019 include/net/sch_generic.h skb->next = *to_free; to_free 1020 include/net/sch_generic.h *to_free = skb; to_free 1024 include/net/sch_generic.h struct sk_buff **to_free) to_free 1027 include/net/sch_generic.h skb->prev->next = *to_free; to_free 1029 include/net/sch_generic.h skb->next = *to_free; to_free 1030 include/net/sch_generic.h *to_free = skb; to_free 1035 include/net/sch_generic.h struct sk_buff **to_free) to_free 1043 include/net/sch_generic.h __qdisc_drop(skb, to_free); to_free 1051 include/net/sch_generic.h struct sk_buff **to_free) to_free 1053 include/net/sch_generic.h return __qdisc_queue_drop_head(sch, &sch->q, to_free); to_free 1174 include/net/sch_generic.h struct sk_buff **to_free) to_free 1176 include/net/sch_generic.h __qdisc_drop(skb, to_free); to_free 1183 include/net/sch_generic.h struct sk_buff **to_free) to_free 1185 include/net/sch_generic.h __qdisc_drop(skb, to_free); to_free 1192 include/net/sch_generic.h struct sk_buff **to_free) to_free 1194 include/net/sch_generic.h __qdisc_drop_all(skb, to_free); to_free 218 mm/kasan/quarantine.c struct qlist_head to_free = QLIST_INIT; to_free 251 mm/kasan/quarantine.c qlist_move_all(&global_quarantine[quarantine_head], &to_free); to_free 252 mm/kasan/quarantine.c WRITE_ONCE(quarantine_size, quarantine_size - to_free.bytes); to_free 260 mm/kasan/quarantine.c qlist_free_all(&to_free, NULL); to_free 291 mm/kasan/quarantine.c struct qlist_head to_free = QLIST_INIT; to_free 295 mm/kasan/quarantine.c qlist_move_cache(q, &to_free, cache); to_free 296 mm/kasan/quarantine.c qlist_free_all(&to_free, cache); to_free 303 mm/kasan/quarantine.c struct qlist_head to_free = QLIST_INIT; to_free 318 mm/kasan/quarantine.c qlist_move_cache(&global_quarantine[i], &to_free, cache); to_free 326 mm/kasan/quarantine.c qlist_free_all(&to_free, cache); to_free 1836 mm/percpu.c LIST_HEAD(to_free); to_free 1855 mm/percpu.c list_move(&chunk->list, &to_free); to_free 1860 mm/percpu.c list_for_each_entry_safe(chunk, next, &to_free, list) { to_free 3383 net/core/dev.c struct sk_buff *to_free = NULL; to_free 3390 net/core/dev.c rc = q->enqueue(skb, q, &to_free) & NET_XMIT_MASK; to_free 3393 net/core/dev.c if (unlikely(to_free)) to_free 3394 net/core/dev.c kfree_skb_list(to_free); to_free 3410 net/core/dev.c __qdisc_drop(skb, &to_free); to_free 3433 net/core/dev.c rc = q->enqueue(skb, q, &to_free) & NET_XMIT_MASK; to_free 3444 net/core/dev.c if (unlikely(to_free)) to_free 3445 net/core/dev.c kfree_skb_list(to_free); to_free 83 net/rds/ib_rdma.c struct rds_ib_ipaddr *to_free = NULL; to_free 90 net/rds/ib_rdma.c to_free = i_ipaddr; to_free 96 net/rds/ib_rdma.c if (to_free) to_free 97 net/rds/ib_rdma.c kfree_rcu(to_free, rcu); to_free 378 net/sched/sch_atm.c struct sk_buff **to_free) to_free 420 net/sched/sch_atm.c __qdisc_drop(skb, to_free); to_free 423 net/sched/sch_atm.c __qdisc_drop(skb, to_free); to_free 435 net/sched/sch_atm.c ret = qdisc_enqueue(skb, flow->q, to_free); to_free 17 net/sched/sch_blackhole.c struct sk_buff **to_free) to_free 19 net/sched/sch_blackhole.c qdisc_drop(skb, sch, to_free); to_free 1461 net/sched/sch_cake.c static unsigned int cake_drop(struct Qdisc *sch, struct sk_buff **to_free) to_free 1510 net/sched/sch_cake.c __qdisc_drop(skb, to_free); to_free 1631 net/sched/sch_cake.c struct sk_buff **to_free) to_free 1647 net/sched/sch_cake.c __qdisc_drop(skb, to_free); to_free 1684 net/sched/sch_cake.c return qdisc_drop(skb, sch, to_free); to_free 1841 net/sched/sch_cake.c cake_drop(sch, to_free); to_free 360 net/sched/sch_cbq.c struct sk_buff **to_free) to_free 372 net/sched/sch_cbq.c __qdisc_drop(skb, to_free); to_free 376 net/sched/sch_cbq.c ret = qdisc_enqueue(skb, cl->q, to_free); to_free 81 net/sched/sch_cbs.c struct sk_buff **to_free); to_free 89 net/sched/sch_cbs.c struct sk_buff **to_free) to_free 94 net/sched/sch_cbs.c err = child->ops->enqueue(skb, child, to_free); to_free 105 net/sched/sch_cbs.c struct sk_buff **to_free) to_free 110 net/sched/sch_cbs.c return cbs_child_enqueue(skb, sch, qdisc, to_free); to_free 114 net/sched/sch_cbs.c struct sk_buff **to_free) to_free 127 net/sched/sch_cbs.c return cbs_child_enqueue(skb, sch, qdisc, to_free); to_free 131 net/sched/sch_cbs.c struct sk_buff **to_free) to_free 135 net/sched/sch_cbs.c return q->enqueue(skb, sch, to_free); to_free 115 net/sched/sch_choke.c struct sk_buff **to_free) to_free 129 net/sched/sch_choke.c qdisc_drop(skb, sch, to_free); to_free 220 net/sched/sch_choke.c struct sk_buff **to_free) to_free 240 net/sched/sch_choke.c choke_drop_by_idx(sch, idx, to_free); to_free 283 net/sched/sch_choke.c return qdisc_drop(skb, sch, to_free); to_free 286 net/sched/sch_choke.c qdisc_drop(skb, sch, to_free); to_free 112 net/sched/sch_codel.c struct sk_buff **to_free) to_free 122 net/sched/sch_codel.c return qdisc_drop(skb, sch, to_free); to_free 341 net/sched/sch_drr.c struct sk_buff **to_free) to_free 353 net/sched/sch_drr.c __qdisc_drop(skb, to_free); to_free 358 net/sched/sch_drr.c err = qdisc_enqueue(skb, cl->qdisc, to_free); to_free 202 net/sched/sch_dsmark.c struct sk_buff **to_free) to_free 253 net/sched/sch_dsmark.c __qdisc_drop(skb, to_free); to_free 270 net/sched/sch_dsmark.c err = qdisc_enqueue(skb, p->q, to_free); to_free 283 net/sched/sch_dsmark.c qdisc_drop(skb, sch, to_free); to_free 163 net/sched/sch_etf.c struct sk_buff **to_free) to_free 173 net/sched/sch_etf.c return qdisc_drop(nskb, sch, to_free); to_free 204 net/sched/sch_etf.c struct sk_buff *to_free = NULL; to_free 223 net/sched/sch_etf.c qdisc_drop(skb, sch, &to_free); to_free 228 net/sched/sch_etf.c kfree_skb_list(to_free); to_free 19 net/sched/sch_fifo.c struct sk_buff **to_free) to_free 24 net/sched/sch_fifo.c return qdisc_drop(skb, sch, to_free); to_free 28 net/sched/sch_fifo.c struct sk_buff **to_free) to_free 33 net/sched/sch_fifo.c return qdisc_drop(skb, sch, to_free); to_free 37 net/sched/sch_fifo.c struct sk_buff **to_free) to_free 46 net/sched/sch_fifo.c __qdisc_queue_drop_head(sch, &sch->q, to_free); to_free 421 net/sched/sch_fq.c struct sk_buff **to_free) to_free 427 net/sched/sch_fq.c return qdisc_drop(skb, sch, to_free); to_free 432 net/sched/sch_fq.c return qdisc_drop(skb, sch, to_free); to_free 139 net/sched/sch_fq_codel.c struct sk_buff **to_free) to_free 172 net/sched/sch_fq_codel.c __qdisc_drop(skb, to_free); to_free 186 net/sched/sch_fq_codel.c struct sk_buff **to_free) to_free 199 net/sched/sch_fq_codel.c __qdisc_drop(skb, to_free); to_free 231 net/sched/sch_fq_codel.c ret = fq_codel_drop(sch, q->drop_batch_size, to_free); to_free 528 net/sched/sch_generic.c struct sk_buff **to_free) to_free 530 net/sched/sch_generic.c __qdisc_drop(skb, to_free); to_free 622 net/sched/sch_generic.c struct sk_buff **to_free) to_free 634 net/sched/sch_generic.c return qdisc_drop_cpu(skb, qdisc, to_free); to_free 636 net/sched/sch_generic.c return qdisc_drop(skb, qdisc, to_free); to_free 165 net/sched/sch_gred.c struct sk_buff **to_free) to_free 253 net/sched/sch_gred.c return qdisc_drop(skb, sch, to_free); to_free 256 net/sched/sch_gred.c qdisc_drop(skb, sch, to_free); to_free 1532 net/sched/sch_hfsc.c hfsc_enqueue(struct sk_buff *skb, struct Qdisc *sch, struct sk_buff **to_free) to_free 1543 net/sched/sch_hfsc.c __qdisc_drop(skb, to_free); to_free 1548 net/sched/sch_hfsc.c err = qdisc_enqueue(skb, cl->qdisc, to_free); to_free 349 net/sched/sch_hhf.c static unsigned int hhf_drop(struct Qdisc *sch, struct sk_buff **to_free) to_free 364 net/sched/sch_hhf.c qdisc_drop(skb, sch, to_free); to_free 372 net/sched/sch_hhf.c struct sk_buff **to_free) to_free 410 net/sched/sch_hhf.c if (hhf_drop(sch, to_free) == idx) to_free 580 net/sched/sch_htb.c struct sk_buff **to_free) to_free 593 net/sched/sch_htb.c return qdisc_drop(skb, sch, to_free); to_free 599 net/sched/sch_htb.c __qdisc_drop(skb, to_free); to_free 603 net/sched/sch_htb.c to_free)) != NET_XMIT_SUCCESS) { to_free 61 net/sched/sch_multiq.c struct sk_buff **to_free) to_free 72 net/sched/sch_multiq.c __qdisc_drop(skb, to_free); to_free 77 net/sched/sch_multiq.c ret = qdisc_enqueue(skb, qdisc, to_free); to_free 413 net/sched/sch_netem.c struct sk_buff **to_free) to_free 421 net/sched/sch_netem.c qdisc_drop(skb, sch, to_free); to_free 435 net/sched/sch_netem.c struct sk_buff **to_free) to_free 463 net/sched/sch_netem.c __qdisc_drop(skb, to_free); to_free 483 net/sched/sch_netem.c rootq->enqueue(skb2, rootq, to_free); to_free 496 net/sched/sch_netem.c skb = netem_segment(skb, sch, to_free); to_free 511 net/sched/sch_netem.c qdisc_drop(skb, sch, to_free); to_free 523 net/sched/sch_netem.c qdisc_drop_all(skb, sch, to_free); to_free 607 net/sched/sch_netem.c rc = qdisc_enqueue(segs, sch, to_free); to_free 720 net/sched/sch_netem.c struct sk_buff *to_free = NULL; to_free 723 net/sched/sch_netem.c err = qdisc_enqueue(skb, q->qdisc, &to_free); to_free 724 net/sched/sch_netem.c kfree_skb_list(to_free); to_free 152 net/sched/sch_pie.c struct sk_buff **to_free) to_free 186 net/sched/sch_pie.c return qdisc_drop(skb, sch, to_free); to_free 88 net/sched/sch_plug.c struct sk_buff **to_free) to_free 98 net/sched/sch_plug.c return qdisc_drop(skb, sch, to_free); to_free 69 net/sched/sch_prio.c prio_enqueue(struct sk_buff *skb, struct Qdisc *sch, struct sk_buff **to_free) to_free 81 net/sched/sch_prio.c __qdisc_drop(skb, to_free); to_free 86 net/sched/sch_prio.c ret = qdisc_enqueue(skb, qdisc, to_free); to_free 1198 net/sched/sch_qfq.c struct sk_buff **to_free) to_free 1211 net/sched/sch_qfq.c __qdisc_drop(skb, to_free); to_free 1222 net/sched/sch_qfq.c return qdisc_drop(skb, sch, to_free); to_free 1228 net/sched/sch_qfq.c err = qdisc_enqueue(skb, cl->qdisc, to_free); to_free 58 net/sched/sch_red.c struct sk_buff **to_free) to_free 97 net/sched/sch_red.c ret = qdisc_enqueue(skb, child, to_free); to_free 108 net/sched/sch_red.c qdisc_drop(skb, sch, to_free); to_free 280 net/sched/sch_sfb.c struct sk_buff **to_free) to_free 402 net/sched/sch_sfb.c ret = qdisc_enqueue(skb, child, to_free); to_free 414 net/sched/sch_sfb.c qdisc_drop(skb, sch, to_free); to_free 293 net/sched/sch_sfq.c static unsigned int sfq_drop(struct Qdisc *sch, struct sk_buff **to_free) to_free 312 net/sched/sch_sfq.c qdisc_drop(skb, sch, to_free); to_free 346 net/sched/sch_sfq.c sfq_enqueue(struct sk_buff *skb, struct Qdisc *sch, struct sk_buff **to_free) to_free 360 net/sched/sch_sfq.c __qdisc_drop(skb, to_free); to_free 370 net/sched/sch_sfq.c return qdisc_drop(skb, sch, to_free); to_free 427 net/sched/sch_sfq.c return qdisc_drop(skb, sch, to_free); to_free 434 net/sched/sch_sfq.c qdisc_drop(head, sch, to_free); to_free 465 net/sched/sch_sfq.c dropped = sfq_drop(sch, to_free); to_free 630 net/sched/sch_sfq.c struct sk_buff *to_free = NULL; to_free 691 net/sched/sch_sfq.c dropped += sfq_drop(sch, &to_free); to_free 693 net/sched/sch_sfq.c tail = to_free; to_free 696 net/sched/sch_sfq.c rtnl_kfree_skbs(to_free, tail); to_free 69 net/sched/sch_skbprio.c struct sk_buff **to_free) to_free 103 net/sched/sch_skbprio.c return qdisc_drop(skb, sch, to_free); to_free 115 net/sched/sch_skbprio.c qdisc_drop(to_drop, sch, to_free); to_free 414 net/sched/sch_taprio.c struct sk_buff **to_free) to_free 424 net/sched/sch_taprio.c return qdisc_drop(skb, sch, to_free); to_free 428 net/sched/sch_taprio.c return qdisc_drop(skb, sch, to_free); to_free 432 net/sched/sch_taprio.c return qdisc_drop(skb, sch, to_free); to_free 438 net/sched/sch_taprio.c return qdisc_enqueue(skb, child, to_free); to_free 144 net/sched/sch_tbf.c struct sk_buff **to_free) to_free 155 net/sched/sch_tbf.c return qdisc_drop(skb, sch, to_free); to_free 163 net/sched/sch_tbf.c ret = qdisc_enqueue(segs, q->qdisc, to_free); to_free 180 net/sched/sch_tbf.c struct sk_buff **to_free) to_free 189 net/sched/sch_tbf.c return tbf_segment(skb, sch, to_free); to_free 190 net/sched/sch_tbf.c return qdisc_drop(skb, sch, to_free); to_free 192 net/sched/sch_tbf.c ret = qdisc_enqueue(skb, q->qdisc, to_free); to_free 76 net/sched/sch_teql.c teql_enqueue(struct sk_buff *skb, struct Qdisc *sch, struct sk_buff **to_free) to_free 86 net/sched/sch_teql.c return qdisc_drop(skb, sch, to_free); to_free 158 tools/perf/util/ordered-events.c list_add(&oe->buffer->list, &oe->to_free); to_free 365 tools/perf/util/ordered-events.c INIT_LIST_HEAD(&oe->to_free); to_free 390 tools/perf/util/ordered-events.c if (list_empty(&oe->to_free)) to_free 403 tools/perf/util/ordered-events.c list_for_each_entry_safe(buffer, tmp, &oe->to_free, list) { to_free 43 tools/perf/util/ordered-events.h struct list_head to_free;