rbnode 41 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode, rbnode 44 drivers/base/regmap/regcache-rbtree.c *base = rbnode->base_reg; rbnode 45 drivers/base/regmap/regcache-rbtree.c *top = rbnode->base_reg + ((rbnode->blklen - 1) * map->reg_stride); rbnode 49 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode, unsigned int idx) rbnode 51 drivers/base/regmap/regcache-rbtree.c return regcache_get_val(map, rbnode->block, idx); rbnode 55 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode, rbnode 58 drivers/base/regmap/regcache-rbtree.c set_bit(idx, rbnode->cache_present); rbnode 59 drivers/base/regmap/regcache-rbtree.c regcache_set_val(map, rbnode->block, idx, val); rbnode 67 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode; rbnode 70 drivers/base/regmap/regcache-rbtree.c rbnode = rbtree_ctx->cached_rbnode; rbnode 71 drivers/base/regmap/regcache-rbtree.c if (rbnode) { rbnode 72 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_get_base_top_reg(map, rbnode, &base_reg, rbnode 75 drivers/base/regmap/regcache-rbtree.c return rbnode; rbnode 80 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); rbnode 81 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_get_base_top_reg(map, rbnode, &base_reg, rbnode 84 drivers/base/regmap/regcache-rbtree.c rbtree_ctx->cached_rbnode = rbnode; rbnode 85 drivers/base/regmap/regcache-rbtree.c return rbnode; rbnode 97 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode) rbnode 112 drivers/base/regmap/regcache-rbtree.c base_reg = rbnode->base_reg; rbnode 125 drivers/base/regmap/regcache-rbtree.c rb_link_node(&rbnode->node, parent, new); rbnode 126 drivers/base/regmap/regcache-rbtree.c rb_insert_color(&rbnode->node, root); rbnode 245 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode; rbnode 248 drivers/base/regmap/regcache-rbtree.c rbnode = regcache_rbtree_lookup(map, reg); rbnode 249 drivers/base/regmap/regcache-rbtree.c if (rbnode) { rbnode 250 drivers/base/regmap/regcache-rbtree.c reg_tmp = (reg - rbnode->base_reg) / map->reg_stride; rbnode 251 drivers/base/regmap/regcache-rbtree.c if (!test_bit(reg_tmp, rbnode->cache_present)) rbnode 253 drivers/base/regmap/regcache-rbtree.c *value = regcache_rbtree_get_register(map, rbnode, reg_tmp); rbnode 263 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode, rbnode 276 drivers/base/regmap/regcache-rbtree.c offset = (rbnode->base_reg - base_reg) / map->reg_stride; rbnode 278 drivers/base/regmap/regcache-rbtree.c blk = krealloc(rbnode->block, rbnode 284 drivers/base/regmap/regcache-rbtree.c if (BITS_TO_LONGS(blklen) > BITS_TO_LONGS(rbnode->blklen)) { rbnode 285 drivers/base/regmap/regcache-rbtree.c present = krealloc(rbnode->cache_present, rbnode 293 drivers/base/regmap/regcache-rbtree.c memset(present + BITS_TO_LONGS(rbnode->blklen), 0, rbnode 294 drivers/base/regmap/regcache-rbtree.c (BITS_TO_LONGS(blklen) - BITS_TO_LONGS(rbnode->blklen)) rbnode 297 drivers/base/regmap/regcache-rbtree.c present = rbnode->cache_present; rbnode 303 drivers/base/regmap/regcache-rbtree.c blk, rbnode->blklen * map->cache_word_size); rbnode 308 drivers/base/regmap/regcache-rbtree.c rbnode->block = blk; rbnode 309 drivers/base/regmap/regcache-rbtree.c rbnode->blklen = blklen; rbnode 310 drivers/base/regmap/regcache-rbtree.c rbnode->base_reg = base_reg; rbnode 311 drivers/base/regmap/regcache-rbtree.c rbnode->cache_present = present; rbnode 313 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_set_register(map, rbnode, pos, value); rbnode 320 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode; rbnode 324 drivers/base/regmap/regcache-rbtree.c rbnode = kzalloc(sizeof(*rbnode), GFP_KERNEL); rbnode 325 drivers/base/regmap/regcache-rbtree.c if (!rbnode) rbnode 338 drivers/base/regmap/regcache-rbtree.c rbnode->blklen = (range->range_max - range->range_min) / rbnode 340 drivers/base/regmap/regcache-rbtree.c rbnode->base_reg = range->range_min; rbnode 344 drivers/base/regmap/regcache-rbtree.c if (!rbnode->blklen) { rbnode 345 drivers/base/regmap/regcache-rbtree.c rbnode->blklen = 1; rbnode 346 drivers/base/regmap/regcache-rbtree.c rbnode->base_reg = reg; rbnode 349 drivers/base/regmap/regcache-rbtree.c rbnode->block = kmalloc_array(rbnode->blklen, map->cache_word_size, rbnode 351 drivers/base/regmap/regcache-rbtree.c if (!rbnode->block) rbnode 354 drivers/base/regmap/regcache-rbtree.c rbnode->cache_present = kcalloc(BITS_TO_LONGS(rbnode->blklen), rbnode 355 drivers/base/regmap/regcache-rbtree.c sizeof(*rbnode->cache_present), rbnode 357 drivers/base/regmap/regcache-rbtree.c if (!rbnode->cache_present) rbnode 360 drivers/base/regmap/regcache-rbtree.c return rbnode; rbnode 363 drivers/base/regmap/regcache-rbtree.c kfree(rbnode->block); rbnode 365 drivers/base/regmap/regcache-rbtree.c kfree(rbnode); rbnode 373 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode, *rbnode_tmp; rbnode 383 drivers/base/regmap/regcache-rbtree.c rbnode = regcache_rbtree_lookup(map, reg); rbnode 384 drivers/base/regmap/regcache-rbtree.c if (rbnode) { rbnode 385 drivers/base/regmap/regcache-rbtree.c reg_tmp = (reg - rbnode->base_reg) / map->reg_stride; rbnode 386 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_set_register(map, rbnode, reg_tmp, value); rbnode 419 drivers/base/regmap/regcache-rbtree.c rbnode = rbnode_tmp; rbnode 439 drivers/base/regmap/regcache-rbtree.c if (rbnode) { rbnode 440 drivers/base/regmap/regcache-rbtree.c ret = regcache_rbtree_insert_to_block(map, rbnode, rbnode 446 drivers/base/regmap/regcache-rbtree.c rbtree_ctx->cached_rbnode = rbnode; rbnode 453 drivers/base/regmap/regcache-rbtree.c rbnode = regcache_rbtree_node_alloc(map, reg); rbnode 454 drivers/base/regmap/regcache-rbtree.c if (!rbnode) rbnode 456 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_set_register(map, rbnode, rbnode 457 drivers/base/regmap/regcache-rbtree.c reg - rbnode->base_reg, value); rbnode 458 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_insert(map, &rbtree_ctx->root, rbnode); rbnode 459 drivers/base/regmap/regcache-rbtree.c rbtree_ctx->cached_rbnode = rbnode; rbnode 470 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode; rbnode 477 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); rbnode 479 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_get_base_top_reg(map, rbnode, &base_reg, rbnode 494 drivers/base/regmap/regcache-rbtree.c end = rbnode->blklen; rbnode 496 drivers/base/regmap/regcache-rbtree.c ret = regcache_sync_block(map, rbnode->block, rbnode 497 drivers/base/regmap/regcache-rbtree.c rbnode->cache_present, rbnode 498 drivers/base/regmap/regcache-rbtree.c rbnode->base_reg, start, end); rbnode 510 drivers/base/regmap/regcache-rbtree.c struct regcache_rbtree_node *rbnode; rbnode 517 drivers/base/regmap/regcache-rbtree.c rbnode = rb_entry(node, struct regcache_rbtree_node, node); rbnode 519 drivers/base/regmap/regcache-rbtree.c regcache_rbtree_get_base_top_reg(map, rbnode, &base_reg, rbnode 534 drivers/base/regmap/regcache-rbtree.c end = rbnode->blklen; rbnode 536 drivers/base/regmap/regcache-rbtree.c bitmap_clear(rbnode->cache_present, start, end - start); rbnode 131 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *rbnode; rbnode 149 drivers/infiniband/hw/hfi1/mmu_rb.c rbnode = rb_entry(node, struct mmu_rb_node, node); rbnode 152 drivers/infiniband/hw/hfi1/mmu_rb.c list_move(&rbnode->list, &del_list); rbnode 237 drivers/infiniband/hw/hfi1/mmu_rb.c struct mmu_rb_node *rbnode, *ptr; rbnode 245 drivers/infiniband/hw/hfi1/mmu_rb.c list_for_each_entry_safe_reverse(rbnode, ptr, &handler->lru_list, rbnode 247 drivers/infiniband/hw/hfi1/mmu_rb.c if (handler->ops->evict(handler->ops_arg, rbnode, evict_arg, rbnode 249 drivers/infiniband/hw/hfi1/mmu_rb.c __mmu_int_rb_remove(rbnode, &handler->root); rbnode 251 drivers/infiniband/hw/hfi1/mmu_rb.c list_move(&rbnode->list, &del_list); rbnode 259 drivers/infiniband/hw/hfi1/mmu_rb.c rbnode = list_first_entry(&del_list, struct mmu_rb_node, list); rbnode 260 drivers/infiniband/hw/hfi1/mmu_rb.c list_del(&rbnode->list); rbnode 261 drivers/infiniband/hw/hfi1/mmu_rb.c handler->ops->remove(handler->ops_arg, rbnode); rbnode 240 fs/btrfs/backref.c ref = rb_entry(parent, struct prelim_ref, rbnode); rbnode 276 fs/btrfs/backref.c rb_link_node(&newref->rbnode, parent, p); rbnode 277 fs/btrfs/backref.c rb_insert_color_cached(&newref->rbnode, root, leftmost); rbnode 289 fs/btrfs/backref.c &preftree->root.rb_root, rbnode) rbnode 635 fs/btrfs/backref.c ref = rb_entry(rnode, struct prelim_ref, rbnode); rbnode 642 fs/btrfs/backref.c rb_erase_cached(&ref->rbnode, &preftrees->indirect.root); rbnode 723 fs/btrfs/backref.c ref = rb_entry(node, struct prelim_ref, rbnode); rbnode 1254 fs/btrfs/backref.c ref = rb_entry(node, struct prelim_ref, rbnode); rbnode 1255 fs/btrfs/backref.c node = rb_next(&ref->rbnode); rbnode 67 fs/btrfs/backref.h struct rb_node rbnode; rbnode 701 include/linux/skbuff.h struct rb_node rbnode; /* used in netem, ip4 defrag, and tcp stack */ rbnode 3398 include/linux/skbuff.h #define rb_to_skb(rb) rb_entry_safe(rb, struct sk_buff, rbnode) rbnode 3401 include/linux/skbuff.h #define skb_rb_next(skb) rb_to_skb(rb_next(&(skb)->rbnode)) rbnode 3402 include/linux/skbuff.h #define skb_rb_prev(skb) rb_to_skb(rb_prev(&(skb)->rbnode)) rbnode 1819 include/net/tcp.h rb_erase(&skb->rbnode, &sk->tcp_rtx_queue); rbnode 152 mm/zswap.c struct rb_node rbnode; rbnode 255 mm/zswap.c RB_CLEAR_NODE(&entry->rbnode); rbnode 273 mm/zswap.c entry = rb_entry(node, struct zswap_entry, rbnode); rbnode 296 mm/zswap.c myentry = rb_entry(parent, struct zswap_entry, rbnode); rbnode 306 mm/zswap.c rb_link_node(&entry->rbnode, parent, link); rbnode 307 mm/zswap.c rb_insert_color(&entry->rbnode, root); rbnode 313 mm/zswap.c if (!RB_EMPTY_NODE(&entry->rbnode)) { rbnode 314 mm/zswap.c rb_erase(&entry->rbnode, root); rbnode 315 mm/zswap.c RB_CLEAR_NODE(&entry->rbnode); rbnode 1212 mm/zswap.c rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode) rbnode 3098 net/core/skbuff.c struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); rbnode 3101 net/core/skbuff.c rb_erase(&skb->rbnode, root); rbnode 48 net/ipv4/inet_fragment.c RB_CLEAR_NODE(&skb->rbnode); rbnode 71 net/ipv4/inet_fragment.c rb_link_node(&skb->rbnode, &q->last_run_head->rbnode, rbnode 72 net/ipv4/inet_fragment.c &q->last_run_head->rbnode.rb_right); rbnode 74 net/ipv4/inet_fragment.c rb_link_node(&skb->rbnode, NULL, &q->rb_fragments.rb_node); rbnode 75 net/ipv4/inet_fragment.c rb_insert_color(&skb->rbnode, &q->rb_fragments); rbnode 238 net/ipv4/inet_fragment.c struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); rbnode 241 net/ipv4/inet_fragment.c rb_erase(&skb->rbnode, root); rbnode 398 net/ipv4/inet_fragment.c rb_link_node(&skb->rbnode, parent, rbn); rbnode 399 net/ipv4/inet_fragment.c rb_insert_color(&skb->rbnode, &q->rb_fragments); rbnode 420 net/ipv4/inet_fragment.c if (RB_EMPTY_NODE(&skb->rbnode)) rbnode 423 net/ipv4/inet_fragment.c rb_replace_node(&skb->rbnode, &fp->rbnode, rbnode 429 net/ipv4/inet_fragment.c rb_replace_node(&head->rbnode, &skb->rbnode, rbnode 489 net/ipv4/inet_fragment.c rbn = rb_next(&head->rbnode); rbnode 490 net/ipv4/inet_fragment.c rb_erase(&head->rbnode, &q->rb_fragments); rbnode 514 net/ipv4/inet_fragment.c memset(&fp->rbnode, 0, sizeof(fp->rbnode)); rbnode 554 net/ipv4/inet_fragment.c rb_replace_node(&head->rbnode, &skb->rbnode, rbnode 557 net/ipv4/inet_fragment.c rb_erase(&head->rbnode, &q->rb_fragments); rbnode 558 net/ipv4/inet_fragment.c memset(&head->rbnode, 0, sizeof(head->rbnode)); rbnode 4499 net/ipv4/tcp_input.c rb_erase(&skb->rbnode, &tp->out_of_order_queue); rbnode 4578 net/ipv4/tcp_input.c rb_link_node(&skb->rbnode, NULL, p); rbnode 4579 net/ipv4/tcp_input.c rb_insert_color(&skb->rbnode, &tp->out_of_order_queue); rbnode 4597 net/ipv4/tcp_input.c parent = &tp->ooo_last_skb->rbnode; rbnode 4628 net/ipv4/tcp_input.c rb_replace_node(&skb1->rbnode, &skb->rbnode, rbnode 4646 net/ipv4/tcp_input.c rb_link_node(&skb->rbnode, parent, p); rbnode 4647 net/ipv4/tcp_input.c rb_insert_color(&skb->rbnode, &tp->out_of_order_queue); rbnode 4659 net/ipv4/tcp_input.c rb_erase(&skb1->rbnode, &tp->out_of_order_queue); rbnode 4877 net/ipv4/tcp_input.c rb_erase(&skb->rbnode, root); rbnode 4900 net/ipv4/tcp_input.c rb_link_node(&skb->rbnode, parent, p); rbnode 4901 net/ipv4/tcp_input.c rb_insert_color(&skb->rbnode, root); rbnode 5085 net/ipv4/tcp_input.c node = &tp->ooo_last_skb->rbnode; rbnode 309 net/netfilter/nf_conncount.c struct rb_node **rbnode, *parent; rbnode 319 net/netfilter/nf_conncount.c rbnode = &(root->rb_node); rbnode 320 net/netfilter/nf_conncount.c while (*rbnode) { rbnode 322 net/netfilter/nf_conncount.c rbconn = rb_entry(*rbnode, struct nf_conncount_rb, node); rbnode 324 net/netfilter/nf_conncount.c parent = *rbnode; rbnode 327 net/netfilter/nf_conncount.c rbnode = &((*rbnode)->rb_left); rbnode 329 net/netfilter/nf_conncount.c rbnode = &((*rbnode)->rb_right); rbnode 377 net/netfilter/nf_conncount.c rb_link_node_rcu(&rbconn->node, parent, rbnode); rbnode 188 net/sched/sch_etf.c rb_link_node(&nskb->rbnode, parent, p); rbnode 189 net/sched/sch_etf.c rb_insert_color_cached(&nskb->rbnode, &q->head, leftmost); rbnode 211 net/sched/sch_etf.c rb_erase_cached(&skb->rbnode, &q->head); rbnode 235 net/sched/sch_etf.c rb_erase_cached(&skb->rbnode, &q->head); rbnode 430 net/sched/sch_etf.c rb_erase_cached(&skb->rbnode, &q->head); rbnode 366 net/sched/sch_fq.c rb_erase(&skb->rbnode, &flow->t_root); rbnode 416 net/sched/sch_fq.c rb_link_node(&skb->rbnode, parent, p); rbnode 417 net/sched/sch_fq.c rb_insert_color(&skb->rbnode, &flow->t_root); rbnode 610 net/sched/sch_fq.c rb_erase(&skb->rbnode, &flow->t_root); rbnode 369 net/sched/sch_netem.c rb_erase(&skb->rbnode, &q->t_root); rbnode 402 net/sched/sch_netem.c rb_link_node(&nskb->rbnode, parent, p); rbnode 403 net/sched/sch_netem.c rb_insert_color(&nskb->rbnode, &q->t_root); rbnode 672 net/sched/sch_netem.c rb_erase(&skb->rbnode, &q->t_root);