/linux-4.1.27/kernel/events/ |
D | ring_buffer.c | 22 atomic_set(&handle->rb->poll, POLLIN); in perf_output_wakeup() 38 struct ring_buffer *rb = handle->rb; in perf_output_get_handle() local 41 local_inc(&rb->nest); in perf_output_get_handle() 42 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle() 47 struct ring_buffer *rb = handle->rb; in perf_output_put_handle() local 51 head = local_read(&rb->head); in perf_output_put_handle() 57 if (!local_dec_and_test(&rb->nest)) in perf_output_put_handle() 87 rb->user_page->data_head = head; in perf_output_put_handle() 93 if (unlikely(head != local_read(&rb->head))) { in perf_output_put_handle() 94 local_inc(&rb->nest); in perf_output_put_handle() [all …]
|
D | internal.h | 58 extern void rb_free(struct ring_buffer *rb); 62 struct ring_buffer *rb; in rb_free_rcu() local 64 rb = container_of(rcu_head, struct ring_buffer, rcu_head); in rb_free_rcu() 65 rb_free(rb); in rb_free_rcu() 71 extern int rb_alloc_aux(struct ring_buffer *rb, struct perf_event *event, 73 extern void rb_free_aux(struct ring_buffer *rb); 75 extern void ring_buffer_put(struct ring_buffer *rb); 77 static inline bool rb_has_aux(struct ring_buffer *rb) in rb_has_aux() argument 79 return !!rb->aux_nr_pages; in rb_has_aux() 95 perf_mmap_to_page(struct ring_buffer *rb, unsigned long pgoff); [all …]
|
D | core.c | 3449 struct ring_buffer *rb); 3597 if (event->rb) { in _free_event() 3917 struct ring_buffer *rb; in perf_poll() local 3930 rb = event->rb; in perf_poll() 3931 if (rb) in perf_poll() 3932 events = atomic_xchg(&rb->poll, 0); in perf_poll() 4241 struct ring_buffer *rb; in perf_event_init_userpage() local 4244 rb = rcu_dereference(event->rb); in perf_event_init_userpage() 4245 if (!rb) in perf_event_init_userpage() 4248 userpg = rb->user_page; in perf_event_init_userpage() [all …]
|
/linux-4.1.27/drivers/scsi/bfa/ |
D | bfa_ioc_ct.c | 192 void __iomem *rb; in bfa_ioc_ct_reg_init() local 195 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 197 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 198 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 199 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 202 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 203 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 204 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() 205 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init() 206 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init() [all …]
|
D | bfa_ioc_cb.c | 145 void __iomem *rb; in bfa_ioc_cb_reg_init() local 148 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init() 150 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init() 151 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init() 152 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init() 155 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init() 156 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init() 157 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init() 159 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init() 160 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init() [all …]
|
D | bfad_debugfs.c | 259 void __iomem *rb, *reg_addr; in bfad_debugfs_write_regrd() local 289 rb = bfa_ioc_bar0(ioc); in bfad_debugfs_write_regrd() 303 reg_addr = rb + addr; in bfad_debugfs_write_regrd()
|
D | bfa_ioc.h | 334 bfa_status_t (*ioc_pll_init) (void __iomem *rb, enum bfi_asic_mode m); 860 bfa_status_t bfa_ioc_cb_pll_init(void __iomem *rb, enum bfi_asic_mode mode); 861 bfa_status_t bfa_ioc_ct_pll_init(void __iomem *rb, enum bfi_asic_mode mode); 862 bfa_status_t bfa_ioc_ct2_pll_init(void __iomem *rb, enum bfi_asic_mode mode);
|
D | bfa_ioc.c | 5520 void __iomem *rb; in bfa_phy_busy() local 5522 rb = bfa_ioc_bar0(ioc); in bfa_phy_busy() 5523 return readl(rb + BFA_PHY_LOCK_STATUS); in bfa_phy_busy()
|
/linux-4.1.27/drivers/net/ethernet/brocade/bna/ |
D | bfa_ioc_ct.c | 58 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb, 60 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb, 260 void __iomem *rb; in bfa_ioc_ct_reg_init() local 263 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 265 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 266 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 267 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 270 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 271 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 272 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() [all …]
|
D | bnad_debugfs.c | 320 void __iomem *rb, *reg_addr; in bnad_debugfs_write_regrd() local 352 rb = bfa_ioc_bar0(ioc); in bnad_debugfs_write_regrd() 366 reg_addr = rb + addr; in bnad_debugfs_write_regrd()
|
D | bfa_ioc.h | 203 enum bfa_status (*ioc_pll_init) (void __iomem *rb,
|
/linux-4.1.27/lib/ |
D | rbtree_test.c | 12 struct rb_node rb; member 31 if (key < rb_entry(parent, struct test_node, rb)->key) in insert() 37 rb_link_node(&node->rb, parent, new); in insert() 38 rb_insert_color(&node->rb, root); in insert() 43 rb_erase(&node->rb, root); in erase() 49 if (node->rb.rb_left) { in augment_recompute() 50 child_augmented = rb_entry(node->rb.rb_left, struct test_node, in augment_recompute() 51 rb)->augmented; in augment_recompute() 55 if (node->rb.rb_right) { in augment_recompute() 56 child_augmented = rb_entry(node->rb.rb_right, struct test_node, in augment_recompute() [all …]
|
D | interval_tree.c | 9 INTERVAL_TREE_DEFINE(struct interval_tree_node, rb,
|
D | rbtree.c | 47 static inline void rb_set_black(struct rb_node *rb) in rb_set_black() argument 49 rb->__rb_parent_color |= RB_BLACK; in rb_set_black()
|
/linux-4.1.27/mm/ |
D | interval_tree.c | 24 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb, 39 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after() 41 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after() 43 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after() 44 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 47 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after() 48 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after() 49 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 53 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after() 57 rb_link_node(&node->shared.rb, &parent->shared.rb, link); in vma_interval_tree_insert_after() [all …]
|
D | kmemleak.c | 407 struct rb_node *rb = object_tree_root.rb_node; in lookup_object() local 409 while (rb) { in lookup_object() 411 rb_entry(rb, struct kmemleak_object, rb_node); in lookup_object() 413 rb = object->rb_node.rb_left; in lookup_object() 415 rb = object->rb_node.rb_right; in lookup_object()
|
D | nommu.c | 1284 struct rb_node *rb; in do_mmap_pgoff() local 1348 for (rb = rb_first(&nommu_region_tree); rb; rb = rb_next(rb)) { in do_mmap_pgoff() 1349 pregion = rb_entry(rb, struct vm_region, vm_rb); in do_mmap_pgoff()
|
/linux-4.1.27/drivers/block/drbd/ |
D | drbd_interval.c | 11 struct drbd_interval *this = rb_entry(node, struct drbd_interval, rb); in interval_end() 27 if (node->rb.rb_left) { in compute_subtree_last() 28 sector_t left = interval_end(node->rb.rb_left); in compute_subtree_last() 32 if (node->rb.rb_right) { in compute_subtree_last() 33 sector_t right = interval_end(node->rb.rb_right); in compute_subtree_last() 40 RB_DECLARE_CALLBACKS(static, augment_callbacks, struct drbd_interval, rb, 56 rb_entry(*new, struct drbd_interval, rb); in drbd_insert_interval() 74 rb_link_node(&this->rb, parent, new); in drbd_insert_interval() 75 rb_insert_augmented(&this->rb, root, &augment_callbacks); in drbd_insert_interval() 97 rb_entry(node, struct drbd_interval, rb); in drbd_contains_interval() [all …]
|
D | drbd_interval.h | 8 struct rb_node rb; member 20 RB_CLEAR_NODE(&i->rb); in drbd_clear_interval() 25 return RB_EMPTY_NODE(&i->rb); in drbd_interval_empty()
|
/linux-4.1.27/fs/jffs2/ |
D | nodelist.h | 230 struct rb_node rb; member 271 struct rb_node rb; member 334 return rb_entry(node, struct jffs2_node_frag, rb); in frag_first() 344 return rb_entry(node, struct jffs2_node_frag, rb); in frag_last() 347 #define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb) 348 #define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb) 349 #define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb) 350 #define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb) 351 #define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb) 352 #define frag_erase(frag, list) rb_erase(&frag->rb, list); [all …]
|
D | nodelist.c | 124 struct rb_node *parent = &base->rb; in jffs2_fragtree_insert() 131 base = rb_entry(parent, struct jffs2_node_frag, rb); in jffs2_fragtree_insert() 134 link = &base->rb.rb_right; in jffs2_fragtree_insert() 136 link = &base->rb.rb_left; in jffs2_fragtree_insert() 143 rb_link_node(&newfrag->rb, &base->rb, link); in jffs2_fragtree_insert() 189 rb_link_node(&holefrag->rb, &this->rb, &this->rb.rb_right); in no_overlapping_node() 193 rb_link_node(&holefrag->rb, NULL, &root->rb_node); in no_overlapping_node() 195 rb_insert_color(&holefrag->rb, root); in no_overlapping_node() 204 rb_link_node(&newfrag->rb, &this->rb, &this->rb.rb_right); in no_overlapping_node() 207 rb_link_node(&newfrag->rb, NULL, &root->rb_node); in no_overlapping_node() [all …]
|
D | readinode.c | 185 tn = rb_entry(next, struct jffs2_tmp_dnode_info, rb); in jffs2_lookup_tn() 188 next = tn->rb.rb_right; in jffs2_lookup_tn() 190 next = tn->rb.rb_left; in jffs2_lookup_tn() 286 rb_replace_node(&this->rb, &tn->rb, &rii->tn_root); in jffs2_add_tn_to_tree() 344 insert_point = rb_entry(parent, struct jffs2_tmp_dnode_info, rb); in jffs2_add_tn_to_tree() 346 link = &insert_point->rb.rb_right; in jffs2_add_tn_to_tree() 349 link = &insert_point->rb.rb_left; in jffs2_add_tn_to_tree() 351 link = &insert_point->rb.rb_right; in jffs2_add_tn_to_tree() 353 rb_link_node(&tn->rb, &insert_point->rb, link); in jffs2_add_tn_to_tree() 354 rb_insert_color(&tn->rb, &rii->tn_root); in jffs2_add_tn_to_tree() [all …]
|
/linux-4.1.27/include/linux/ |
D | rbtree_augmented.h | 66 rbname ## _propagate(struct rb_node *rb, struct rb_node *stop) \ 68 while (rb != stop) { \ 69 rbstruct *node = rb_entry(rb, rbstruct, rbfield); \ 74 rb = rb_parent(&node->rbfield); \ 105 #define rb_color(rb) __rb_color((rb)->__rb_parent_color) argument 106 #define rb_is_red(rb) __rb_is_red((rb)->__rb_parent_color) argument 107 #define rb_is_black(rb) __rb_is_black((rb)->__rb_parent_color) argument 109 static inline void rb_set_parent(struct rb_node *rb, struct rb_node *p) in rb_set_parent() argument 111 rb->__rb_parent_color = rb_color(rb) | (unsigned long)p; in rb_set_parent() 114 static inline void rb_set_parent_color(struct rb_node *rb, in rb_set_parent_color() argument [all …]
|
D | interval_tree_generic.h | 158 struct rb_node *rb = node->ITRB.rb_right, *prev; \ 168 if (rb) { \ 169 ITSTRUCT *right = rb_entry(rb, ITSTRUCT, ITRB); \ 177 rb = rb_parent(&node->ITRB); \ 178 if (!rb) \ 181 node = rb_entry(rb, ITSTRUCT, ITRB); \ 182 rb = node->ITRB.rb_right; \ 183 } while (prev == rb); \
|
D | interval_tree.h | 7 struct rb_node rb; member
|
D | rmap.h | 77 struct rb_node rb; /* locked by anon_vma->rwsem */ member
|
D | mm_types.h | 278 struct rb_node rb; member
|
D | perf_event.h | 454 struct ring_buffer *rb; member 573 struct ring_buffer *rb; member
|
D | kernfs.h | 121 struct rb_node rb; member
|
D | fb.h | 719 extern int fb_find_mode_cvt(struct fb_videomode *mode, int margins, int rb);
|
/linux-4.1.27/arch/x86/mm/ |
D | pat_rbtree.c | 51 struct memtype *data = container_of(node, struct memtype, rb); in get_subtree_max_end() 61 child_max_end = get_subtree_max_end(data->rb.rb_right); in compute_subtree_max_end() 65 child_max_end = get_subtree_max_end(data->rb.rb_left); in compute_subtree_max_end() 72 RB_DECLARE_CALLBACKS(static, memtype_rb_augment_cb, struct memtype, rb, in RB_DECLARE_CALLBACKS() argument 83 struct memtype *data = container_of(node, struct memtype, rb); in RB_DECLARE_CALLBACKS() 113 node = rb_next(&match->rb); in memtype_rb_exact_match() 115 match = container_of(node, struct memtype, rb); in memtype_rb_exact_match() 142 node = rb_next(&match->rb); in memtype_rb_check_conflict() 144 match = container_of(node, struct memtype, rb); in memtype_rb_check_conflict() 154 node = rb_next(&match->rb); in memtype_rb_check_conflict() [all …]
|
D | pat_internal.h | 14 struct rb_node rb; member
|
/linux-4.1.27/drivers/xen/xenbus/ |
D | xenbus_dev_frontend.c | 126 struct read_buffer *rb; in xenbus_file_read() local 144 rb = list_entry(u->read_buffers.next, struct read_buffer, list); in xenbus_file_read() 147 unsigned sz = min((unsigned)len - i, rb->len - rb->cons); in xenbus_file_read() 149 ret = copy_to_user(ubuf + i, &rb->msg[rb->cons], sz); in xenbus_file_read() 152 rb->cons += sz - ret; in xenbus_file_read() 161 if (rb->cons == rb->len) { in xenbus_file_read() 162 list_del(&rb->list); in xenbus_file_read() 163 kfree(rb); in xenbus_file_read() 166 rb = list_entry(u->read_buffers.next, in xenbus_file_read() 187 struct read_buffer *rb; in queue_reply() local [all …]
|
/linux-4.1.27/arch/arm/lib/ |
D | getuser.S | 45 rb .req ip label 47 3: ldrbt rb, [r0], #0 49 rb .req r0 label 51 3: ldrb rb, [r0, #1] 54 orr r2, r2, rb, lsl #8 56 orr r2, rb, r2, lsl #8 105 rb .req ip label 107 10: ldrbt rb, [r0], #0 109 rb .req r0 label 111 10: ldrb rb, [r0, #1] [all …]
|
/linux-4.1.27/drivers/target/iscsi/ |
D | iscsi_target_configfs.c | 56 ssize_t rb; in lio_target_np_show_sctp() local 60 rb = sprintf(page, "1\n"); in lio_target_np_show_sctp() 62 rb = sprintf(page, "0\n"); in lio_target_np_show_sctp() 64 return rb; in lio_target_np_show_sctp() 132 ssize_t rb; in lio_target_np_show_iser() local 136 rb = sprintf(page, "1\n"); in lio_target_np_show_iser() 138 rb = sprintf(page, "0\n"); in lio_target_np_show_iser() 140 return rb; in lio_target_np_show_iser() 581 ssize_t rb; \ 586 rb = snprintf(page, PAGE_SIZE, \ [all …]
|
/linux-4.1.27/arch/tile/kernel/ |
D | unaligned.c | 180 uint64_t *rb, uint64_t *clob1, uint64_t *clob2, in find_regs() argument 208 *rb = reg; in find_regs() 209 alias_reg_map = (1ULL << *ra) | (1ULL << *rb); in find_regs() 260 *rb = reg; in find_regs() 261 alias_reg_map = (1ULL << *ra) | (1ULL << *rb); in find_regs() 316 static bool check_regs(uint64_t rd, uint64_t ra, uint64_t rb, in check_regs() argument 330 if ((rb >= 56) && (rb != TREG_ZERO)) in check_regs() 394 static tilegx_bundle_bits jit_x0_dblalign(int rd, int ra, int rb) in jit_x0_dblalign() argument 399 create_SrcB_X0(rb); in jit_x0_dblalign() 436 static tilegx_bundle_bits jit_x1_st1_add(int ra, int rb, int imm8) in jit_x1_st1_add() argument [all …]
|
/linux-4.1.27/arch/sparc/kernel/ |
D | btext.c | 22 static void draw_byte_32(unsigned char *bits, unsigned int *base, int rb); 23 static void draw_byte_16(unsigned char *bits, unsigned int *base, int rb); 24 static void draw_byte_8(unsigned char *bits, unsigned int *base, int rb); 197 int rb = dispDeviceRowBytes; in draw_byte() local 202 draw_byte_32(font, (unsigned int *)base, rb); in draw_byte() 206 draw_byte_16(font, (unsigned int *)base, rb); in draw_byte() 209 draw_byte_8(font, (unsigned int *)base, rb); in draw_byte() 241 static void draw_byte_32(unsigned char *font, unsigned int *base, int rb) in draw_byte_32() argument 258 base = (unsigned int *) ((char *)base + rb); in draw_byte_32() 262 static void draw_byte_16(unsigned char *font, unsigned int *base, int rb) in draw_byte_16() argument [all …]
|
/linux-4.1.27/tools/perf/ |
D | builtin-lock.c | 43 struct rb_node rb; /* used for sorting */ member 110 struct rb_node rb; member 125 st = container_of(node, struct thread_stat, rb); in thread_stat_find() 139 struct rb_node **rb = &thread_stats.rb_node; in thread_stat_insert() local 143 while (*rb) { in thread_stat_insert() 144 p = container_of(*rb, struct thread_stat, rb); in thread_stat_insert() 145 parent = *rb; in thread_stat_insert() 148 rb = &(*rb)->rb_left; in thread_stat_insert() 150 rb = &(*rb)->rb_right; in thread_stat_insert() 155 rb_link_node(&new->rb, parent, rb); in thread_stat_insert() [all …]
|
D | builtin-kvm.c | 480 struct rb_node **rb = &result->rb_node; in insert_to_result() local 484 while (*rb) { in insert_to_result() 485 p = container_of(*rb, struct kvm_event, rb); in insert_to_result() 486 parent = *rb; in insert_to_result() 489 rb = &(*rb)->rb_left; in insert_to_result() 491 rb = &(*rb)->rb_right; in insert_to_result() 494 rb_link_node(&event->rb, parent, rb); in insert_to_result() 495 rb_insert_color(&event->rb, result); in insert_to_result() 538 return container_of(node, struct kvm_event, rb); in pop_from_result()
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | kvm_book3s_64.h | 137 unsigned long rb = 0, va_low, sllp; in compute_tlbie_rb() local 162 rb = (v & ~0x7fUL) << 16; /* AVA field */ in compute_tlbie_rb() 164 rb |= (v >> HPTE_V_SSIZE_SHIFT) << 8; /* B field */ in compute_tlbie_rb() 188 rb |= sllp << 5; /* AP field */ in compute_tlbie_rb() 189 rb |= (va_low & 0x7ff) << 12; /* remaining 11 bits of AVA */ in compute_tlbie_rb() 198 rb |= (va_low << mmu_psize_defs[b_psize].shift) & 0x7ff000; in compute_tlbie_rb() 202 rb &= ~((1ul << mmu_psize_defs[a_psize].shift) - 1); in compute_tlbie_rb() 209 rb |= ((va_low << aval_shift) & 0xfe); in compute_tlbie_rb() 211 rb |= 1; /* L field */ in compute_tlbie_rb() 213 rb |= penc << 12; /* LP field */ in compute_tlbie_rb() [all …]
|
D | ppc_asm.h | 27 #define ACCOUNT_CPU_USER_ENTRY(ra, rb) argument 28 #define ACCOUNT_CPU_USER_EXIT(ra, rb) argument 31 #define ACCOUNT_CPU_USER_ENTRY(ra, rb) \ argument 33 ld rb,PACA_STARTTIME_USER(r13); \ 35 subf rb,rb,ra; /* subtract start value */ \ 37 add ra,ra,rb; /* add on to user time */ \ 40 #define ACCOUNT_CPU_USER_EXIT(ra, rb) \ argument 42 ld rb,PACA_STARTTIME(r13); \ 44 subf rb,rb,ra; /* subtract start value */ \ 46 add ra,ra,rb; /* add on to system time */ \ [all …]
|
D | asm-compat.h | 76 #define PPC405_ERR77(ra,rb) stringify_in_c(dcbt ra, rb;) argument 79 #define PPC405_ERR77(ra,rb) argument
|
D | exception-64s.h | 119 #define SAVE_PPR(area, ra, rb) \ argument 122 ld rb,area+EX_PPR(r13); /* Read PPR from paca */ \ 123 std rb,TASKTHREADPPR(ra); \
|
D | kvm_ppc.h | 663 static inline ulong kvmppc_get_ea_indexed(struct kvm_vcpu *vcpu, int ra, int rb) in kvmppc_get_ea_indexed() argument 668 ea = kvmppc_get_gpr(vcpu, rb); in kvmppc_get_ea_indexed()
|
D | kvm_host.h | 337 void (*slbmte)(struct kvm_vcpu *vcpu, u64 rb, u64 rs);
|
/linux-4.1.27/fs/ocfs2/ |
D | refcounttree.c | 84 struct ocfs2_refcount_block *rb = in ocfs2_validate_refcount_block() local 96 rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &rb->rf_check); in ocfs2_validate_refcount_block() 104 if (!OCFS2_IS_VALID_REFCOUNT_BLOCK(rb)) { in ocfs2_validate_refcount_block() 108 rb->rf_signature); in ocfs2_validate_refcount_block() 112 if (le64_to_cpu(rb->rf_blkno) != bh->b_blocknr) { in ocfs2_validate_refcount_block() 117 (unsigned long long)le64_to_cpu(rb->rf_blkno)); in ocfs2_validate_refcount_block() 121 if (le32_to_cpu(rb->rf_fs_generation) != OCFS2_SB(sb)->fs_generation) { in ocfs2_validate_refcount_block() 126 le32_to_cpu(rb->rf_fs_generation)); in ocfs2_validate_refcount_block() 460 struct ocfs2_refcount_block *rb; in ocfs2_lock_refcount_tree() local 487 rb = (struct ocfs2_refcount_block *)ref_root_bh->b_data; in ocfs2_lock_refcount_tree() [all …]
|
D | xattr.c | 6283 struct ocfs2_refcount_block *rb = in ocfs2_reflink_lock_xattr_allocators() local 6307 if (le32_to_cpu(rb->rf_flags) & OCFS2_REFCOUNT_TREE_FL) in ocfs2_reflink_lock_xattr_allocators() 6308 *credits += le16_to_cpu(rb->rf_list.l_tree_depth) * in ocfs2_reflink_lock_xattr_allocators() 6309 le16_to_cpu(rb->rf_list.l_next_free_rec) + 1; in ocfs2_reflink_lock_xattr_allocators() 6751 struct ocfs2_refcount_block *rb; in ocfs2_lock_reflink_xattr_rec_allocators() local 6775 rb = (struct ocfs2_refcount_block *)args->reflink->ref_root_bh->b_data; in ocfs2_lock_reflink_xattr_rec_allocators() 6782 if (le32_to_cpu(rb->rf_flags) & OCFS2_REFCOUNT_TREE_FL) in ocfs2_lock_reflink_xattr_rec_allocators() 6783 *credits += le16_to_cpu(rb->rf_list.l_tree_depth) * in ocfs2_lock_reflink_xattr_rec_allocators() 6784 le16_to_cpu(rb->rf_list.l_next_free_rec) + 1; in ocfs2_lock_reflink_xattr_rec_allocators()
|
D | alloc.c | 392 struct ocfs2_refcount_block *rb = et->et_object; in ocfs2_refcount_tree_fill_root_el() local 394 et->et_root_el = &rb->rf_list; in ocfs2_refcount_tree_fill_root_el() 400 struct ocfs2_refcount_block *rb = et->et_object; in ocfs2_refcount_tree_set_last_eb_blk() local 402 rb->rf_last_eb_blk = cpu_to_le64(blkno); in ocfs2_refcount_tree_set_last_eb_blk() 407 struct ocfs2_refcount_block *rb = et->et_object; in ocfs2_refcount_tree_get_last_eb_blk() local 409 return le64_to_cpu(rb->rf_last_eb_blk); in ocfs2_refcount_tree_get_last_eb_blk() 415 struct ocfs2_refcount_block *rb = et->et_object; in ocfs2_refcount_tree_update_clusters() local 417 le32_add_cpu(&rb->rf_clusters, clusters); in ocfs2_refcount_tree_update_clusters()
|
/linux-4.1.27/drivers/mtd/ubi/ |
D | attach.c | 270 av = rb_entry(parent, struct ubi_ainf_volume, rb); in add_volume() 297 rb_link_node(&av->rb, parent, p); in add_volume() 298 rb_insert_color(&av->rb, &ai->volumes); in add_volume() 481 aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); in ubi_add_to_av() 592 rb_link_node(&aeb->u.rb, parent, p); in ubi_add_to_av() 593 rb_insert_color(&aeb->u.rb, &av->root); in ubi_add_to_av() 612 av = rb_entry(p, struct ubi_ainf_volume, rb); in ubi_find_av() 633 struct rb_node *rb; in ubi_remove_av() local 638 while ((rb = rb_first(&av->root))) { in ubi_remove_av() 639 aeb = rb_entry(rb, struct ubi_ainf_peb, u.rb); in ubi_remove_av() [all …]
|
D | wl.c | 161 e1 = rb_entry(parent, struct ubi_wl_entry, u.rb); in wl_tree_add() 176 rb_link_node(&e->u.rb, parent, p); in wl_tree_add() 177 rb_insert_color(&e->u.rb, root); in wl_tree_add() 257 e1 = rb_entry(p, struct ubi_wl_entry, u.rb); in in_wl_tree() 317 e = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); in find_wl_entry() 324 e1 = rb_entry(p, struct ubi_wl_entry, u.rb); in find_wl_entry() 358 first = rb_entry(rb_first(root), struct ubi_wl_entry, u.rb); in find_mean_wl_entry() 359 last = rb_entry(rb_last(root), struct ubi_wl_entry, u.rb); in find_mean_wl_entry() 362 e = rb_entry(root->rb_node, struct ubi_wl_entry, u.rb); in find_mean_wl_entry() 398 rb_erase(&e->u.rb, &ubi->free); in wl_get_wle() [all …]
|
D | ubi.h | 178 struct rb_node rb; member 200 struct rb_node rb; member 655 struct rb_node rb; member 690 struct rb_node rb; member 941 ubi_rb_for_each_entry((tmp_rb), (e), &(ubi)->free, u.rb) 950 ubi_rb_for_each_entry((tmp_rb), (e), &(ubi)->used, u.rb) 959 ubi_rb_for_each_entry((tmp_rb), (e), &(ubi)->scrub, u.rb) 978 #define ubi_rb_for_each_entry(rb, pos, root, member) \ argument 979 for (rb = rb_first(root), \ 980 pos = (rb ? container_of(rb, typeof(*pos), member) : NULL); \ [all …]
|
D | fastmap-wl.c | 41 ubi_rb_for_each_entry(p, e, root, u.rb) { in find_anchor_wl_entry() 74 ubi_rb_for_each_entry(p, e, root, u.rb) in anchor_pebs_avalible() 109 rb_erase(&e->u.rb, &ubi->free); in ubi_wl_get_fm_peb() 156 rb_erase(&e->u.rb, &ubi->free); in ubi_refill_pools() 359 struct ubi_wl_entry, u.rb); in may_reserve_for_fm()
|
D | eba.c | 105 le = rb_entry(p, struct ubi_ltree_entry, rb); in ltree_lookup() 171 le1 = rb_entry(parent, struct ubi_ltree_entry, rb); in ltree_add_entry() 186 rb_link_node(&le->rb, parent, p); in ltree_add_entry() 187 rb_insert_color(&le->rb, &ubi->ltree); in ltree_add_entry() 232 rb_erase(&le->rb, &ubi->ltree); in leb_read_unlock() 284 rb_erase(&le->rb, &ubi->ltree); in leb_write_trylock() 308 rb_erase(&le->rb, &ubi->ltree); in leb_write_unlock() 1321 struct rb_node *rb; in self_check_eba() local 1361 ubi_rb_for_each_entry(rb, aeb, &av->root, u.rb) in self_check_eba() 1368 ubi_rb_for_each_entry(rb, aeb, &av->root, u.rb) in self_check_eba() [all …]
|
D | fastmap.c | 191 av = rb_entry(parent, struct ubi_ainf_volume, rb); in add_vol() 215 rb_link_node(&av->rb, parent, p); in add_vol() 216 rb_insert_color(&av->rb, &ai->volumes); in add_vol() 240 tmp_aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); in assign_aeb_to_av() 255 rb_link_node(&aeb->u.rb, parent, p); in assign_aeb_to_av() 256 rb_insert_color(&aeb->u.rb, &av->root); in assign_aeb_to_av() 279 aeb = rb_entry(parent, struct ubi_ainf_peb, u.rb); in update_vol() 350 rb_link_node(&new_aeb->u.rb, parent, p); in update_vol() 351 rb_insert_color(&new_aeb->u.rb, &av->root); in update_vol() 383 tmp_av = rb_entry(parent, struct ubi_ainf_volume, rb); in process_pool_aeb() [all …]
|
D | vtbl.c | 373 struct rb_node *rb; in process_lvol() local 406 ubi_rb_for_each_entry(rb, aeb, &av->root, u.rb) { in process_lvol()
|
/linux-4.1.27/arch/powerpc/lib/ |
D | sstep.c | 131 int ra, rb; in xform_ea() local 135 rb = (instr >> 11) & 0x1f; in xform_ea() 136 ea = regs->gpr[rb]; in xform_ea() 645 unsigned int opcode, ra, rb, rd, spr, u; in analyse_instr() local 728 rb = (instr >> 11) & 0x1f; in analyse_instr() 731 rb = (regs->ccr >> (31 - rb)) & 1; in analyse_instr() 732 val = (instr >> (6 + ra * 2 + rb)) & 1; in analyse_instr() 769 rb = (instr >> 11) & 0x1f; in analyse_instr() 842 regs->gpr[ra] = (regs->gpr[ra] & ~imm) | (ROTATE(val, rb) & imm); in analyse_instr() 849 regs->gpr[ra] = ROTATE(val, rb) & MASK32(mb, me); in analyse_instr() [all …]
|
/linux-4.1.27/net/sunrpc/xprtrdma/ |
D | xprt_rdma.h | 128 rdmab_addr(struct rpcrdma_regbuf *rb) in rdmab_addr() argument 130 return rb->rg_iov.addr; in rdmab_addr() 134 rdmab_length(struct rpcrdma_regbuf *rb) in rdmab_length() argument 136 return rb->rg_iov.length; in rdmab_length() 140 rdmab_lkey(struct rpcrdma_regbuf *rb) in rdmab_lkey() argument 142 return rb->rg_iov.lkey; in rdmab_lkey() 146 rdmab_to_msg(struct rpcrdma_regbuf *rb) in rdmab_to_msg() argument 148 return (struct rpcrdma_msg *)rb->rg_base; in rdmab_to_msg() 271 struct rpcrdma_regbuf *rb; in rpcr_to_rdmar() local 273 rb = container_of(buffer, struct rpcrdma_regbuf, rg_base); in rpcr_to_rdmar() [all …]
|
D | transport.c | 495 struct rpcrdma_regbuf *rb; in xprt_rdma_allocate() local 522 rb = rpcrdma_alloc_regbuf(&r_xprt->rx_ia, min_size, flags); in xprt_rdma_allocate() 523 if (IS_ERR(rb)) in xprt_rdma_allocate() 525 req->rl_rdmabuf = rb; in xprt_rdma_allocate() 549 rb = rpcrdma_alloc_regbuf(&r_xprt->rx_ia, size, flags); in xprt_rdma_allocate() 550 if (IS_ERR(rb)) in xprt_rdma_allocate() 552 rb->rg_owner = req; in xprt_rdma_allocate() 556 req->rl_sendbuf = rb; in xprt_rdma_allocate() 573 struct rpcrdma_regbuf *rb; in xprt_rdma_free() local 579 rb = container_of(buffer, struct rpcrdma_regbuf, rg_base[0]); in xprt_rdma_free() [all …]
|
D | verbs.c | 1534 struct rpcrdma_regbuf *rb; in rpcrdma_alloc_regbuf() local 1538 rb = kmalloc(sizeof(*rb) + size, flags); in rpcrdma_alloc_regbuf() 1539 if (rb == NULL) in rpcrdma_alloc_regbuf() 1542 rb->rg_size = size; in rpcrdma_alloc_regbuf() 1543 rb->rg_owner = NULL; in rpcrdma_alloc_regbuf() 1544 rc = rpcrdma_register_internal(ia, rb->rg_base, size, in rpcrdma_alloc_regbuf() 1545 &rb->rg_mr, &rb->rg_iov); in rpcrdma_alloc_regbuf() 1549 return rb; in rpcrdma_alloc_regbuf() 1552 kfree(rb); in rpcrdma_alloc_regbuf() 1563 rpcrdma_free_regbuf(struct rpcrdma_ia *ia, struct rpcrdma_regbuf *rb) in rpcrdma_free_regbuf() argument [all …]
|
/linux-4.1.27/drivers/tty/hvc/ |
D | hvc_iucv.c | 221 struct iucv_tty_buffer *rb; in hvc_iucv_write() local 239 rb = list_first_entry(&priv->tty_inqueue, struct iucv_tty_buffer, list); in hvc_iucv_write() 242 if (!rb->mbuf) { /* message not yet received ... */ in hvc_iucv_write() 245 rb->mbuf = kmalloc(rb->msg.length, GFP_ATOMIC | GFP_DMA); in hvc_iucv_write() 246 if (!rb->mbuf) in hvc_iucv_write() 249 rc = __iucv_message_receive(priv->path, &rb->msg, 0, in hvc_iucv_write() 250 rb->mbuf, rb->msg.length, NULL); in hvc_iucv_write() 262 if (rc || (rb->mbuf->version != MSG_VERSION) || in hvc_iucv_write() 263 (rb->msg.length != MSG_SIZE(rb->mbuf->datalen))) in hvc_iucv_write() 267 switch (rb->mbuf->type) { in hvc_iucv_write() [all …]
|
/linux-4.1.27/fs/nilfs2/ |
D | recovery.c | 353 struct nilfs_recovery_block *rb; in nilfs_scan_dsync_log() local 361 rb = kmalloc(sizeof(*rb), GFP_NOFS); in nilfs_scan_dsync_log() 362 if (unlikely(!rb)) { in nilfs_scan_dsync_log() 366 rb->ino = ino; in nilfs_scan_dsync_log() 367 rb->blocknr = blocknr++; in nilfs_scan_dsync_log() 368 rb->vblocknr = le64_to_cpu(binfo->bi_vblocknr); in nilfs_scan_dsync_log() 369 rb->blkoff = le64_to_cpu(binfo->bi_blkoff); in nilfs_scan_dsync_log() 371 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log() 390 struct nilfs_recovery_block *rb; in dispose_recovery_list() local 392 rb = list_first_entry(head, struct nilfs_recovery_block, list); in dispose_recovery_list() [all …]
|
/linux-4.1.27/arch/powerpc/kvm/ |
D | book3s_pr_papr.c | 90 unsigned long v = 0, pteg, rb; in kvmppc_h_pr_remove() local 108 rb = compute_tlbie_rb(pte[0], pte[1], pte_index); in kvmppc_h_pr_remove() 109 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_remove() 150 unsigned long pteg, rb, flags; in kvmppc_h_pr_bulk_remove() local 189 rb = compute_tlbie_rb(pte[0], pte[1], in kvmppc_h_pr_bulk_remove() 191 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_bulk_remove() 208 unsigned long rb, pteg, r, v; in kvmppc_h_pr_protect() local 233 rb = compute_tlbie_rb(v, r, pte_index); in kvmppc_h_pr_protect() 234 vcpu->arch.mmu.tlbie(vcpu, rb, rb & 1 ? true : false); in kvmppc_h_pr_protect()
|
D | e500_emulate.c | 52 static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb) in kvmppc_e500_emul_msgclr() argument 54 ulong param = vcpu->arch.gpr[rb]; in kvmppc_e500_emul_msgclr() 64 static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb) in kvmppc_e500_emul_msgsnd() argument 66 ulong param = vcpu->arch.gpr[rb]; in kvmppc_e500_emul_msgsnd() 67 int prio = dbell2prio(rb); in kvmppc_e500_emul_msgsnd() 121 int rb = get_rb(inst); in kvmppc_core_emulate_op_e500() local 135 emulated = kvmppc_e500_emul_msgsnd(vcpu, rb); in kvmppc_core_emulate_op_e500() 139 emulated = kvmppc_e500_emul_msgclr(vcpu, rb); in kvmppc_core_emulate_op_e500() 152 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); in kvmppc_core_emulate_op_e500() 158 ea = kvmppc_get_ea_indexed(vcpu, ra, rb); in kvmppc_core_emulate_op_e500() [all …]
|
D | book3s_hv_ras.c | 54 unsigned long rb = be64_to_cpu(slb->save_area[i].esid); in reload_slb() local 57 rb = (rb & ~0xFFFul) | i; /* insert entry number */ in reload_slb() 58 asm volatile("slbmte %0,%1" : : "r" (rs), "r" (rb)); in reload_slb()
|
D | book3s_64_mmu.c | 378 static void kvmppc_mmu_book3s_64_slbmte(struct kvm_vcpu *vcpu, u64 rs, u64 rb) in kvmppc_mmu_book3s_64_slbmte() argument 385 dprintk("KVM MMU: slbmte(0x%llx, 0x%llx)\n", rs, rb); in kvmppc_mmu_book3s_64_slbmte() 389 esid = GET_ESID(rb); in kvmppc_mmu_book3s_64_slbmte() 390 esid_1t = GET_ESID_1T(rb); in kvmppc_mmu_book3s_64_slbmte() 391 slb_nr = rb & 0xfff; in kvmppc_mmu_book3s_64_slbmte() 402 slbe->valid = (rb & SLB_ESID_V) ? 1 : 0; in kvmppc_mmu_book3s_64_slbmte() 423 slbe->orige = rb & (ESID_MASK | SLB_ESID_V); in kvmppc_mmu_book3s_64_slbmte() 497 u64 rb = 0, rs = 0; in kvmppc_mmu_book3s_64_mtsrin() local 520 rb |= (srnum & 0xf) << 28; in kvmppc_mmu_book3s_64_mtsrin() 522 rb |= 1 << 27; in kvmppc_mmu_book3s_64_mtsrin() [all …]
|
D | book3s_emulate.c | 96 int rb = get_rb(inst); in kvmppc_core_emulate_op_pr() local 166 srnum = (kvmppc_get_gpr(vcpu, rb) >> 28) & 0xf; in kvmppc_core_emulate_op_pr() 181 (kvmppc_get_gpr(vcpu, rb) >> 28) & 0xf, in kvmppc_core_emulate_op_pr() 188 ulong addr = kvmppc_get_gpr(vcpu, rb); in kvmppc_core_emulate_op_pr() 228 kvmppc_get_gpr(vcpu, rb)); in kvmppc_core_emulate_op_pr() 235 kvmppc_get_gpr(vcpu, rb)); in kvmppc_core_emulate_op_pr() 249 rb_val = kvmppc_get_gpr(vcpu, rb); in kvmppc_core_emulate_op_pr() 260 rb_val = kvmppc_get_gpr(vcpu, rb); in kvmppc_core_emulate_op_pr() 270 ulong rb_val = kvmppc_get_gpr(vcpu, rb); in kvmppc_core_emulate_op_pr() 673 ulong rb = get_rb(inst); in kvmppc_alignment_dar() [all …]
|
D | book3s_hv_rm_mmu.c | 404 unsigned long v, r, rb; in kvmppc_do_h_remove() local 425 rb = compute_tlbie_rb(v, be64_to_cpu(hpte[1]), pte_index); in kvmppc_do_h_remove() 426 do_tlbies(kvm, &rb, 1, global_invalidates(kvm, flags), true); in kvmppc_do_h_remove() 573 unsigned long v, r, rb, mask, bits; in kvmppc_h_protect() local 617 rb = compute_tlbie_rb(v, r, pte_index); in kvmppc_h_protect() 620 do_tlbies(kvm, &rb, 1, global_invalidates(kvm, flags), in kvmppc_h_protect() 667 unsigned long rb; in kvmppc_invalidate_hpte() local 670 rb = compute_tlbie_rb(be64_to_cpu(hptep[0]), be64_to_cpu(hptep[1]), in kvmppc_invalidate_hpte() 672 do_tlbies(kvm, &rb, 1, 1, true); in kvmppc_invalidate_hpte() 679 unsigned long rb; in kvmppc_clear_ref_hpte() local [all …]
|
/linux-4.1.27/Documentation/ |
D | rbtree.txt | 246 Interval tree is an example of augmented rb tree. Reference - 272 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 275 if (node->rb.rb_left) { 277 rb_entry(node->rb.rb_left, 278 struct interval_tree_node, rb); 295 if (node->rb.rb_right) { 296 node = rb_entry(node->rb.rb_right, 297 struct interval_tree_node, rb); 312 if (node->rb.rb_left) { 313 subtree_last = rb_entry(node->rb.rb_left, [all …]
|
D | kprobes.txt | 649 r = unoptimized kretprobe, rb = boosted kretprobe, ro = optimized kretprobe. 652 k = 0.80 usec; b = 0.33; o = 0.05; r = 1.10; rb = 0.61; ro = 0.33 655 k = 0.99 usec; b = 0.43; o = 0.06; r = 1.24; rb = 0.68; ro = 0.30
|
D | java.txt | 294 classfile = fopen(argv[1], "rb");
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | mce_power.c | 33 unsigned long rb; in flush_tlb_206() local 38 rb = TLBIEL_INVAL_SET; in flush_tlb_206() 41 rb = TLBIEL_INVAL_SET_LPID; in flush_tlb_206() 50 asm volatile("tlbiel %0" : : "r" (rb)); in flush_tlb_206() 51 rb += 1 << TLBIEL_INVAL_SET_SHIFT; in flush_tlb_206() 107 unsigned long rb = be64_to_cpu(slb->save_area[i].esid); in flush_and_reload_slb() local 110 rb = (rb & ~0xFFFul) | i; in flush_and_reload_slb() 111 asm volatile("slbmte %0,%1" : : "r" (rs), "r" (rb)); in flush_and_reload_slb()
|
D | btext.c | 425 static void draw_byte_32(unsigned char *font, unsigned int *base, int rb) in draw_byte_32() argument 442 base = (unsigned int *) ((char *)base + rb); in draw_byte_32() 446 static inline void draw_byte_16(unsigned char *font, unsigned int *base, int rb) in draw_byte_16() argument 460 base = (unsigned int *) ((char *)base + rb); in draw_byte_16() 464 static inline void draw_byte_8(unsigned char *font, unsigned int *base, int rb) in draw_byte_8() argument 476 base = (unsigned int *) ((char *)base + rb); in draw_byte_8() 484 int rb = dispDeviceRowBytes; in draw_byte() local 490 draw_byte_32(font, (unsigned int *)base, rb); in draw_byte() 494 draw_byte_16(font, (unsigned int *)base, rb); in draw_byte() 497 draw_byte_8(font, (unsigned int *)base, rb); in draw_byte()
|
D | time.c | 1082 u64 ra, rb, rc; in div128_by_32() local 1092 rb = ((u64) do_div(ra, divisor) << 32) + c; in div128_by_32() 1095 rc = ((u64) do_div(rb, divisor) << 32) + d; in div128_by_32() 1096 y = rb; in div128_by_32()
|
D | kvm.c | 376 static void kvm_patch_ins_mtsrin(u32 *inst, u32 rt, u32 rb) in kvm_patch_ins_mtsrin() argument 401 p[kvm_emulate_mtsrin_reg1_offs] |= (rb << 10); in kvm_patch_ins_mtsrin()
|
D | traps.c | 350 unsigned int rb; in check_io_access() local 353 rb = (*nip >> 11) & 0x1f; in check_io_access() 356 regs->gpr[rb] - _IO_BASE, nip); in check_io_access()
|
/linux-4.1.27/drivers/i2c/ |
D | i2c-stub.c | 101 struct smbus_block_data *b, *rb = NULL; in stub_find_block() local 105 rb = b; in stub_find_block() 109 if (rb == NULL && create) { in stub_find_block() 110 rb = devm_kzalloc(dev, sizeof(*rb), GFP_KERNEL); in stub_find_block() 111 if (rb == NULL) in stub_find_block() 112 return rb; in stub_find_block() 113 rb->command = command; in stub_find_block() 114 list_add(&rb->node, &chip->smbus_blocks); in stub_find_block() 116 return rb; in stub_find_block()
|
/linux-4.1.27/drivers/gpu/drm/msm/adreno/ |
D | adreno_gpu.c | 60 ret = msm_gem_get_iova(gpu->rb->bo, gpu->id, &gpu->rb_iova); in adreno_hw_init() 70 AXXX_CP_RB_CNTL_BUFSZ(ilog2(gpu->rb->size / 8)) | in adreno_hw_init() 107 gpu->rb->cur = gpu->rb->start; in adreno_recover() 127 struct msm_ringbuffer *ring = gpu->rb; in adreno_submit() 196 uint32_t wptr = get_wptr(gpu->rb); in adreno_flush() 207 uint32_t wptr = get_wptr(gpu->rb); in adreno_idle() 231 seq_printf(m, "rb wptr: %d\n", get_wptr(gpu->rb)); in adreno_show() 267 printk("rb wptr: %d\n", get_wptr(gpu->rb)); in adreno_dump() 286 uint32_t size = gpu->rb->size / 4; in ring_freewords() 287 uint32_t wptr = get_wptr(gpu->rb); in ring_freewords()
|
D | a4xx_gpu.c | 112 struct msm_ringbuffer *ring = gpu->rb; in a4xx_me_init()
|
D | a3xx_gpu.c | 46 struct msm_ringbuffer *ring = gpu->rb; in a3xx_me_init()
|
/linux-4.1.27/fs/ubifs/ |
D | orphan.c | 85 o = rb_entry(parent, struct ubifs_orphan, rb); in ubifs_add_orphan() 99 rb_link_node(&orphan->rb, parent, p); in ubifs_add_orphan() 100 rb_insert_color(&orphan->rb, &c->orph_tree); in ubifs_add_orphan() 123 o = rb_entry(p, struct ubifs_orphan, rb); in ubifs_delete_orphan() 451 rb_erase(&orphan->rb, &c->orph_tree); in erase_deleted() 525 o = rb_entry(parent, struct ubifs_orphan, rb); in insert_dead_orphan() 537 rb_link_node(&orphan->rb, parent, p); in insert_dead_orphan() 538 rb_insert_color(&orphan->rb, &c->orph_tree); in insert_dead_orphan() 734 struct rb_node rb; member 755 o = rb_entry(p, struct ubifs_orphan, rb); in dbg_find_orphan() [all …]
|
D | log.c | 50 bud = rb_entry(p, struct ubifs_bud, rb); in ubifs_search_bud() 83 bud = rb_entry(p, struct ubifs_bud, rb); in ubifs_get_wbuf() 134 b = rb_entry(parent, struct ubifs_bud, rb); in ubifs_add_bud() 142 rb_link_node(&bud->rb, parent, p); in ubifs_add_bud() 143 rb_insert_color(&bud->rb, &c->buds); in ubifs_add_bud() 314 bud = rb_entry(p1, struct ubifs_bud, rb); in remove_buds() 540 struct rb_node rb; member 559 dr = rb_entry(parent, struct done_ref, rb); in done_already() 574 rb_link_node(&dr->rb, parent, p); in done_already() 575 rb_insert_color(&dr->rb, done_tree); in done_already() [all …]
|
D | debug.c | 580 struct rb_node *rb; in ubifs_dump_budg() local 622 for (rb = rb_first(&c->buds); rb; rb = rb_next(rb)) { in ubifs_dump_budg() 623 bud = rb_entry(rb, struct ubifs_bud, rb); in ubifs_dump_budg() 648 struct rb_node *rb; in ubifs_dump_lprop() local 710 for (rb = rb_first((struct rb_root *)&c->buds); rb; rb = rb_next(rb)) { in ubifs_dump_lprop() 711 bud = rb_entry(rb, struct ubifs_bud, rb); in ubifs_dump_lprop() 1760 struct rb_node rb; member 1807 fscki = rb_entry(parent, struct fsck_inode, rb); in add_inode() 1863 rb_link_node(&fscki->rb, parent, p); in add_inode() 1864 rb_insert_color(&fscki->rb, &fsckd->inodes); in add_inode() [all …]
|
D | recovery.c | 1242 struct rb_node rb; member 1266 e = rb_entry(parent, struct size_entry, rb); in add_ino() 1282 rb_link_node(&e->rb, parent, p); in add_ino() 1283 rb_insert_color(&e->rb, &c->size_tree); in add_ino() 1299 e = rb_entry(p, struct size_entry, rb); in find_ino() 1321 rb_erase(&e->rb, &c->size_tree); in remove_ino() 1333 rbtree_postorder_for_each_entry_safe(e, n, &c->size_tree, rb) { in ubifs_destroy_size_tree() 1483 e = rb_entry(this, struct size_entry, rb); in ubifs_recover_size() 1542 rb_erase(&e->rb, &c->size_tree); in ubifs_recover_size()
|
D | replay.c | 279 struct replay_entry *ra, *rb; in replay_entries_cmp() local 286 rb = list_entry(b, struct replay_entry, list); in replay_entries_cmp() 287 ubifs_assert(ra->sqnum != rb->sqnum); in replay_entries_cmp() 288 if (ra->sqnum > rb->sqnum) in replay_entries_cmp()
|
D | ubifs.h | 283 struct rb_node rb; member 723 struct rb_node rb; member 923 struct rb_node rb; member
|
D | tnc.c | 91 o = rb_entry(parent, struct ubifs_old_idx, rb); in insert_old_idx() 106 rb_link_node(&old_idx->rb, parent, p); in insert_old_idx() 107 rb_insert_color(&old_idx->rb, &c->old_idx); in insert_old_idx() 183 rbtree_postorder_for_each_entry_safe(old_idx, n, &c->old_idx, rb) in destroy_old_idx()
|
D | tnc_commit.c | 175 o = rb_entry(p, struct ubifs_old_idx, rb); in find_old_idx()
|
D | super.c | 875 rbtree_postorder_for_each_entry_safe(bud, n, &c->buds, rb) in free_buds()
|
/linux-4.1.27/net/packet/ |
D | af_packet.c | 194 struct packet_ring_buffer *rb, 489 struct packet_ring_buffer *rb, in packet_lookup_frame() argument 496 pg_vec_pos = position / rb->frames_per_block; in packet_lookup_frame() 497 frame_offset = position % rb->frames_per_block; in packet_lookup_frame() 499 h.raw = rb->pg_vec[pg_vec_pos].buffer + in packet_lookup_frame() 500 (frame_offset * rb->frame_size); in packet_lookup_frame() 509 struct packet_ring_buffer *rb, in packet_current_frame() argument 512 return packet_lookup_frame(po, rb, rb->head, status); in packet_current_frame() 608 struct packet_ring_buffer *rb, in init_prb_bdqc() argument 612 struct tpacket_kbdq_core *p1 = GET_PBDQC_FROM_RB(rb); in init_prb_bdqc() [all …]
|
/linux-4.1.27/fs/btrfs/ |
D | extent_map.c | 228 struct rb_node *rb; in try_merge_map() local 231 rb = rb_prev(&em->rb_node); in try_merge_map() 232 if (rb) in try_merge_map() 233 merge = rb_entry(rb, struct extent_map, rb_node); in try_merge_map() 234 if (rb && mergable_maps(merge, em)) { in try_merge_map() 250 rb = rb_next(&em->rb_node); in try_merge_map() 251 if (rb) in try_merge_map() 252 merge = rb_entry(rb, struct extent_map, rb_node); in try_merge_map() 253 if (rb && mergable_maps(em, merge)) { in try_merge_map()
|
D | raid56.c | 1660 struct btrfs_raid_bio *rb = container_of(b, struct btrfs_raid_bio, in plug_cmp() local 1663 u64 b_sector = rb->bio_list.head->bi_iter.bi_sector; in plug_cmp()
|
/linux-4.1.27/tools/testing/selftests/powerpc/primitives/asm/ |
D | asm-compat.h | 76 #define PPC405_ERR77(ra,rb) stringify_in_c(dcbt ra, rb;) argument 79 #define PPC405_ERR77(ra,rb) argument
|
/linux-4.1.27/drivers/staging/lustre/lnet/lnet/ |
D | router.c | 1235 lnet_destroy_rtrbuf(lnet_rtrbuf_t *rb, int npages) in lnet_destroy_rtrbuf() argument 1240 __free_page(rb->rb_kiov[npages].kiov_page); in lnet_destroy_rtrbuf() 1242 LIBCFS_FREE(rb, sz); in lnet_destroy_rtrbuf() 1251 lnet_rtrbuf_t *rb; in lnet_new_rtrbuf() local 1254 LIBCFS_CPT_ALLOC(rb, lnet_cpt_table(), cpt, sz); in lnet_new_rtrbuf() 1255 if (rb == NULL) in lnet_new_rtrbuf() 1258 rb->rb_pool = rbp; in lnet_new_rtrbuf() 1266 __free_page(rb->rb_kiov[i].kiov_page); in lnet_new_rtrbuf() 1268 LIBCFS_FREE(rb, sz); in lnet_new_rtrbuf() 1272 rb->rb_kiov[i].kiov_len = PAGE_CACHE_SIZE; in lnet_new_rtrbuf() [all …]
|
D | lib-move.c | 899 lnet_rtrbuf_t *rb; in lnet_post_routed_recv_locked() local 950 rb = list_entry(rbp->rbp_bufs.next, lnet_rtrbuf_t, rb_list); in lnet_post_routed_recv_locked() 951 list_del(&rb->rb_list); in lnet_post_routed_recv_locked() 954 msg->msg_kiov = &rb->rb_kiov[0]; in lnet_post_routed_recv_locked() 1033 lnet_rtrbuf_t *rb; in lnet_return_rx_credits_locked() local 1041 rb = list_entry(msg->msg_kiov, lnet_rtrbuf_t, rb_kiov[0]); in lnet_return_rx_credits_locked() 1042 rbp = rb->rb_pool; in lnet_return_rx_credits_locked() 1053 list_add(&rb->rb_list, &rbp->rbp_bufs); in lnet_return_rx_credits_locked()
|
/linux-4.1.27/Documentation/devicetree/bindings/mtd/ |
D | sunxi-nand.txt | 18 - allwinner,rb : shall contain the native Ready/Busy ids. 20 - rb-gpios : shall contain the gpios used as R/B pins. 42 allwinner,rb = <0>;
|
/linux-4.1.27/crypto/ |
D | tgr192.c | 401 static void tgr192_round(u64 * ra, u64 * rb, u64 * rc, u64 x, int mul) in tgr192_round() argument 404 u64 b = *rb; in tgr192_round() 415 *rb = b; in tgr192_round() 420 static void tgr192_pass(u64 * ra, u64 * rb, u64 * rc, u64 * x, int mul) in tgr192_pass() argument 423 u64 b = *rb; in tgr192_pass() 436 *rb = b; in tgr192_pass()
|
/linux-4.1.27/fs/kernfs/ |
D | dir.c | 25 #define rb_to_kn(X) rb_entry((X), struct kernfs_node, rb) 247 node = &pos->rb.rb_left; in kernfs_link_sibling() 249 node = &pos->rb.rb_right; in kernfs_link_sibling() 255 rb_link_node(&kn->rb, parent, node); in kernfs_link_sibling() 256 rb_insert_color(&kn->rb, &kn->parent->dir.children); in kernfs_link_sibling() 278 if (RB_EMPTY_NODE(&kn->rb)) in kernfs_unlink_sibling() 284 rb_erase(&kn->rb, &kn->parent->dir.children); in kernfs_unlink_sibling() 285 RB_CLEAR_NODE(&kn->rb); in kernfs_unlink_sibling() 535 RB_CLEAR_NODE(&kn->rb); in __kernfs_new_node() 985 rbn = rb_next(&pos->rb); in kernfs_next_descendant_post() [all …]
|
/linux-4.1.27/drivers/spi/ |
D | spi-mpc52xx-psc.c | 137 unsigned rb = 0; /* number of bytes receieved */ in mpc52xx_psc_spi_transfer_rxtx() local 151 while (rb < t->len) { in mpc52xx_psc_spi_transfer_rxtx() 152 if (t->len - rb > MPC52xx_PSC_BUFSIZE) { in mpc52xx_psc_spi_transfer_rxtx() 157 rfalarm = MPC52xx_PSC_BUFSIZE - (t->len - rb); in mpc52xx_psc_spi_transfer_rxtx() 179 if (t->len - rb == 1) { in mpc52xx_psc_spi_transfer_rxtx() 192 for (; recv_at_once; rb++, recv_at_once--) in mpc52xx_psc_spi_transfer_rxtx() 193 rx_buf[rb] = in_8(&psc->mpc52xx_psc_buffer_8); in mpc52xx_psc_spi_transfer_rxtx() 195 for (; recv_at_once; rb++, recv_at_once--) in mpc52xx_psc_spi_transfer_rxtx()
|
/linux-4.1.27/drivers/i2c/busses/ |
D | i2c-cpm.c | 194 u_char *rb; in cpm_i2c_parse_message() local 205 rb = cpm->rxbuf[rx]; in cpm_i2c_parse_message() 208 rb = (u_char *) (((ulong) rb + 1) & ~1); in cpm_i2c_parse_message() 254 u_char *rb; in cpm_i2c_check_message() local 261 rb = cpm->rxbuf[rx]; in cpm_i2c_check_message() 264 rb = (u_char *) (((uint) rb + 1) & ~1); in cpm_i2c_check_message() 284 memcpy(pmsg->buf, rb, pmsg->len); in cpm_i2c_check_message()
|
/linux-4.1.27/drivers/media/dvb-frontends/ |
D | dib3000mb.c | 56 u8 rb[2]; in dib3000_read_reg() local 59 { .addr = state->config.demod_address, .flags = I2C_M_RD, .buf = rb, .len = 2 }, in dib3000_read_reg() 66 (rb[0] << 8) | rb[1],(rb[0] << 8) | rb[1]); in dib3000_read_reg() 68 return (rb[0] << 8) | rb[1]; in dib3000_read_reg()
|
D | dib3000mc.c | 53 u8 rb[2]; in dib3000mc_read_word() local 56 { .addr = state->i2c_addr >> 1, .flags = I2C_M_RD, .buf = rb, .len = 2 }, in dib3000mc_read_word() 62 return (rb[0] << 8) | rb[1]; in dib3000mc_read_word()
|
/linux-4.1.27/drivers/mtd/nand/ |
D | sunxi_nand.c | 193 struct sunxi_nand_rb rb; member 339 struct sunxi_nand_rb *rb; in sunxi_nfc_dev_ready() local 346 rb = &sunxi_nand->sels[sunxi_nand->selected].rb; in sunxi_nfc_dev_ready() 348 switch (rb->type) { in sunxi_nfc_dev_ready() 351 (NFC_RB_STATE0 << rb->info.nativeid)); in sunxi_nfc_dev_ready() 357 (NFC_RB_STATE0 << rb->info.nativeid)); in sunxi_nfc_dev_ready() 360 ret = gpio_get_value(rb->info.gpio); in sunxi_nfc_dev_ready() 394 if (sel->rb.type == RB_NONE) { in sunxi_nfc_select_chip() 398 if (sel->rb.type == RB_NATIVE) in sunxi_nfc_select_chip() 399 ctl |= (sel->rb.info.nativeid << 3); in sunxi_nfc_select_chip() [all …]
|
/linux-4.1.27/drivers/iio/ |
D | industrialio-buffer.c | 101 struct iio_buffer *rb = indio_dev->buffer; in iio_buffer_read_first_n_outer() local 110 if (!rb || !rb->access->read_first_n) in iio_buffer_read_first_n_outer() 113 datum_size = rb->bytes_per_datum; in iio_buffer_read_first_n_outer() 122 to_read = min_t(size_t, n / datum_size, rb->watermark); in iio_buffer_read_first_n_outer() 128 ret = wait_event_interruptible(rb->pollq, in iio_buffer_read_first_n_outer() 129 iio_buffer_ready(indio_dev, rb, to_wait, to_read)); in iio_buffer_read_first_n_outer() 136 ret = rb->access->read_first_n(rb, n, buf); in iio_buffer_read_first_n_outer() 151 struct iio_buffer *rb = indio_dev->buffer; in iio_buffer_poll() local 156 poll_wait(filp, &rb->pollq, wait); in iio_buffer_poll() 157 if (iio_buffer_ready(indio_dev, rb, rb->watermark, 0)) in iio_buffer_poll()
|
/linux-4.1.27/arch/mips/alchemy/common/ |
D | usb.c | 391 static inline int au1000_usb_init(unsigned long rb, int reg) in au1000_usb_init() argument 393 void __iomem *base = (void __iomem *)KSEG1ADDR(rb + reg); in au1000_usb_init() 424 static inline void __au1xx0_ohci_control(int enable, unsigned long rb, int creg) in __au1xx0_ohci_control() argument 426 void __iomem *base = (void __iomem *)KSEG1ADDR(rb); in __au1xx0_ohci_control() 457 static inline int au1000_usb_control(int block, int enable, unsigned long rb, in au1000_usb_control() argument 464 __au1xx0_ohci_control(enable, rb, creg); in au1000_usb_control()
|
/linux-4.1.27/drivers/gpu/drm/msm/ |
D | msm_gpu.c | 618 gpu->rb = msm_ringbuffer_new(gpu, ringsz); in msm_gpu_init() 620 if (IS_ERR(gpu->rb)) { in msm_gpu_init() 621 ret = PTR_ERR(gpu->rb); in msm_gpu_init() 622 gpu->rb = NULL; in msm_gpu_init() 643 if (gpu->rb) { in msm_gpu_cleanup() 645 msm_gem_put_iova(gpu->rb->bo, gpu->id); in msm_gpu_cleanup() 646 msm_ringbuffer_destroy(gpu->rb); in msm_gpu_cleanup()
|
D | msm_gpu.h | 80 struct msm_ringbuffer *rb; member
|
/linux-4.1.27/arch/powerpc/sysdev/ |
D | mpic.c | 174 struct mpic_reg_bank *rb, in _mpic_read() argument 180 return dcr_read(rb->dhost, reg); in _mpic_read() 183 return in_be32(rb->base + (reg >> 2)); in _mpic_read() 186 return in_le32(rb->base + (reg >> 2)); in _mpic_read() 191 struct mpic_reg_bank *rb, in _mpic_write() argument 197 dcr_write(rb->dhost, reg, value); in _mpic_write() 201 out_be32(rb->base + (reg >> 2), value); in _mpic_write() 205 out_le32(rb->base + (reg >> 2), value); in _mpic_write() 315 struct mpic_reg_bank *rb, unsigned int offset, in _mpic_map_mmio() argument 318 rb->base = ioremap(phys_addr + offset, size); in _mpic_map_mmio() [all …]
|
D | fsl_pci.c | 864 unsigned int rd, ra, rb, d; in mcheck_handle_load() local 868 rb = get_rb(inst); in mcheck_handle_load() 881 regs->gpr[ra] += regs->gpr[rb]; in mcheck_handle_load() 890 regs->gpr[ra] += regs->gpr[rb]; in mcheck_handle_load() 900 regs->gpr[ra] += regs->gpr[rb]; in mcheck_handle_load() 909 regs->gpr[ra] += regs->gpr[rb]; in mcheck_handle_load()
|
/linux-4.1.27/drivers/usb/class/ |
D | cdc-acm.c | 417 struct acm_rb *rb = urb->context; in acm_read_bulk_callback() local 418 struct acm *acm = rb->instance; in acm_read_bulk_callback() 423 rb->index, urb->actual_length); in acm_read_bulk_callback() 426 set_bit(rb->index, &acm->read_urbs_free); in acm_read_bulk_callback() 432 set_bit(rb->index, &acm->read_urbs_free); in acm_read_bulk_callback() 447 set_bit(rb->index, &acm->read_urbs_free); in acm_read_bulk_callback() 454 acm_submit_read_urb(acm, rb->index, GFP_ATOMIC); in acm_read_bulk_callback() 1371 struct acm_rb *rb = &(acm->read_buffers[i]); in acm_probe() local 1374 rb->base = usb_alloc_coherent(acm->dev, readsize, GFP_KERNEL, in acm_probe() 1375 &rb->dma); in acm_probe() [all …]
|
/linux-4.1.27/drivers/mtd/ |
D | mtdswap.c | 83 struct rb_node rb; member 93 rb)->erase_count) 95 rb)->erase_count) 214 rb_erase(&eb->rb, eb->root); in mtdswap_eb_detach() 226 cur = rb_entry(parent, struct swap_eb, rb); in __mtdswap_rb_add() 233 rb_link_node(&eb->rb, parent, p); in __mtdswap_rb_add() 234 rb_insert_color(&eb->rb, root); in __mtdswap_rb_add() 453 median = rb_entry(medrb, struct swap_eb, rb)->erase_count; in mtdswap_check_counts() 466 rb_erase(&eb->rb, &hist_root); in mtdswap_check_counts() 625 eb = rb_entry(rb_first(clean_root), struct swap_eb, rb); in mtdswap_map_free_block() [all …]
|
/linux-4.1.27/arch/arm/crypto/ |
D | sha512-armv7-neon.S | 84 #define rounds2_0_63(ra, rb, rc, rd, re, rf, rg, rh, rw0, rw1, rw01q, rw2, \ argument 114 veor.64 RT0, ra, rb; \ 116 vbsl.64 RT0, rc, rb; \ 151 vbsl.64 RT0, rb, ra; \ 195 #define rounds2_64_79(ra, rb, rc, rd, re, rf, rg, rh, rw0, rw1, \ argument 226 veor.64 RT0, ra, rb; \ 228 vbsl.64 RT0, rc, rb; \ 263 vbsl.64 RT0, rb, ra; \
|
/linux-4.1.27/fs/xfs/ |
D | xfs_rtalloc.h | 127 # define xfs_rtallocate_extent(t,b,min,max,l,a,f,p,rb) (ENOSYS) argument 129 # define xfs_rtpick_extent(m,t,l,rb) (ENOSYS) argument
|
/linux-4.1.27/drivers/media/platform/exynos4-is/ |
D | fimc-isp-video.c | 534 struct v4l2_requestbuffers *rb) in isp_video_reqbufs() argument 539 ret = vb2_ioctl_reqbufs(file, priv, rb); in isp_video_reqbufs() 543 if (rb->count && rb->count < FIMC_ISP_REQ_BUFS_MIN) { in isp_video_reqbufs() 544 rb->count = 0; in isp_video_reqbufs() 545 vb2_ioctl_reqbufs(file, priv, rb); in isp_video_reqbufs() 549 isp->video_capture.reqbufs_count = rb->count; in isp_video_reqbufs()
|
/linux-4.1.27/drivers/media/pci/ngene/ |
D | ngene-core.c | 777 static void free_ringbuffer(struct ngene *dev, struct SRingBufferDescriptor *rb) in free_ringbuffer() argument 779 struct SBufferHeader *Cur = rb->Head; in free_ringbuffer() 785 for (j = 0; j < rb->NumBuffers; j++, Cur = Cur->Next) { in free_ringbuffer() 788 rb->Buffer1Length, in free_ringbuffer() 794 rb->Buffer2Length, in free_ringbuffer() 799 if (rb->SCListMem) in free_ringbuffer() 800 pci_free_consistent(dev->pci_dev, rb->SCListMemSize, in free_ringbuffer() 801 rb->SCListMem, rb->PASCListMem); in free_ringbuffer() 803 pci_free_consistent(dev->pci_dev, rb->MemSize, rb->Head, rb->PAHead); in free_ringbuffer() 807 struct SRingBufferDescriptor *rb, in free_idlebuffer() argument [all …]
|
/linux-4.1.27/drivers/usb/gadget/function/ |
D | uvc_queue.c | 152 struct v4l2_requestbuffers *rb) in uvcg_alloc_buffers() argument 156 ret = vb2_reqbufs(&queue->queue, rb); in uvcg_alloc_buffers() 158 return ret ? ret : rb->count; in uvcg_alloc_buffers()
|
D | uvc_queue.h | 65 struct v4l2_requestbuffers *rb);
|
/linux-4.1.27/sound/pci/lola/ |
D | lola.c | 367 PAGE_SIZE, &chip->rb); in setup_corb_rirb() 371 chip->corb.addr = chip->rb.addr; in setup_corb_rirb() 372 chip->corb.buf = (u32 *)chip->rb.area; in setup_corb_rirb() 373 chip->rirb.addr = chip->rb.addr + 2048; in setup_corb_rirb() 374 chip->rirb.buf = (u32 *)(chip->rb.area + 2048); in setup_corb_rirb() 556 if (chip->rb.area) in lola_free() 557 snd_dma_free_pages(&chip->rb); in lola_free()
|
D | lola.h | 344 struct snd_dma_buffer rb; member
|
/linux-4.1.27/arch/arm/boot/compressed/ |
D | head.S | 30 .macro loadsp, rb, tmp 32 .macro writeb, ch, rb 36 .macro loadsp, rb, tmp 38 .macro writeb, ch, rb 42 .macro loadsp, rb, tmp 44 .macro writeb, ch, rb 53 .macro writeb, ch, rb 54 senduart \ch, \rb 58 .macro loadsp, rb, tmp 59 mov \rb, #0x80000000 @ physical base address [all …]
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_modes.c | 1231 bool yres_specified = false, cvt = false, rb = false; in drm_mode_parse_command_line_for_connector() local 1252 !yres_specified && !cvt && !rb && was_digit) { in drm_mode_parse_command_line_for_connector() 1261 !rb && was_digit) { in drm_mode_parse_command_line_for_connector() 1285 if (yres_specified || cvt || rb || was_digit) in drm_mode_parse_command_line_for_connector() 1287 rb = true; in drm_mode_parse_command_line_for_connector() 1364 mode->rb = rb; in drm_mode_parse_command_line_for_connector() 1392 cmd->rb, cmd->interlace, in drm_mode_create_from_cmdline_mode()
|
D | drm_edid.c | 539 short rb; member 1471 bool rb) in drm_mode_find_dmt() argument 1483 if (rb != mode_is_rb(ptr)) in drm_mode_find_dmt() 2080 bool rb = drm_monitor_supports_rb(edid); in drm_cvt_modes_for_range() local 2084 newmode = drm_cvt_mode(dev, m->w, m->h, m->r, rb, 0, 0); in drm_cvt_modes_for_range() 2172 est3_modes[m].rb); in drm_est3_modes()
|
D | drm_fb_helper.c | 1296 if (cmdline_mode->rb || cmdline_mode->margins) in drm_pick_cmdline_mode()
|
D | drm_crtc.c | 842 mode->rb ? " reduced blanking" : "", in drm_connector_get_cmdline_mode()
|
/linux-4.1.27/drivers/crypto/vmx/ |
D | ppc-xlate.pl | 144 my ($f, $vrt, $ra, $rb, $op) = @_; 145 " .long ".sprintf "0x%X",(31<<26)|($vrt<<21)|($ra<<16)|($rb<<11)|($op*2+1);
|
/linux-4.1.27/drivers/infiniband/hw/usnic/ |
D | usnic_uiom_interval_tree.h | 25 struct rb_node rb; member
|
D | usnic_uiom_interval_tree.c | 252 INTERVAL_TREE_DEFINE(struct usnic_uiom_interval_node, rb,
|
/linux-4.1.27/drivers/media/usb/gspca/ |
D | gspca.c | 1371 struct v4l2_requestbuffers *rb) in vidioc_reqbufs() argument 1376 i = rb->memory; /* (avoid compilation warning) */ in vidioc_reqbufs() 1390 && gspca_dev->memory != rb->memory) { in vidioc_reqbufs() 1424 if (rb->count == 0) /* unrequest */ in vidioc_reqbufs() 1426 ret = frame_alloc(gspca_dev, file, rb->memory, rb->count); in vidioc_reqbufs() 1428 rb->count = gspca_dev->nframes; in vidioc_reqbufs() 1434 PDEBUG(D_STREAM, "reqbufs st:%d c:%d", ret, rb->count); in vidioc_reqbufs() 1823 struct v4l2_requestbuffers rb; in read_alloc() local 1825 memset(&rb, 0, sizeof rb); in read_alloc() 1826 rb.count = gspca_dev->nbufread; in read_alloc() [all …]
|
/linux-4.1.27/drivers/misc/ |
D | sram.c | 52 struct sram_reserve *rb = list_entry(b, struct sram_reserve, list); in sram_reserve_cmp() local 54 return ra->start - rb->start; in sram_reserve_cmp()
|
/linux-4.1.27/drivers/media/usb/uvc/ |
D | uvc_queue.c | 228 struct v4l2_requestbuffers *rb) in uvc_request_buffers() argument 233 ret = vb2_reqbufs(&queue->queue, rb); in uvc_request_buffers() 236 return ret ? ret : rb->count; in uvc_request_buffers()
|
D | uvcvideo.h | 631 struct v4l2_requestbuffers *rb);
|
D | uvc_v4l2.c | 681 struct v4l2_requestbuffers *rb) in uvc_ioctl_reqbufs() argument 692 ret = uvc_request_buffers(&stream->queue, rb); in uvc_ioctl_reqbufs()
|
/linux-4.1.27/drivers/video/fbdev/core/ |
D | modedb.c | 700 int yres_specified = 0, cvt = 0, rb = 0, interlace = 0; in fb_find_mode() local 713 if (cvt || rb) in fb_find_mode() 724 if (cvt || rb) in fb_find_mode() 745 rb = 1; in fb_find_mode() 770 (rb) ? " reduced blanking" : "", in fb_find_mode() 784 ret = fb_find_mode_cvt(&cvt_mode, margins, rb); in fb_find_mode()
|
D | fbcvt.c | 305 int fb_find_mode_cvt(struct fb_videomode *mode, int margins, int rb) in fb_find_mode_cvt() argument 314 if (rb) in fb_find_mode_cvt()
|
/linux-4.1.27/drivers/infiniband/core/ |
D | umem_rbtree.c | 69 INTERVAL_TREE_DEFINE(struct umem_odp_node, rb, u64, __subtree_last,
|
/linux-4.1.27/drivers/media/platform/s3c-camif/ |
D | camif-capture.c | 908 struct v4l2_requestbuffers *rb) in s3c_camif_reqbufs() argument 914 vp->id, rb->count, vp->owner, priv); in s3c_camif_reqbufs() 919 if (rb->count) in s3c_camif_reqbufs() 920 rb->count = max_t(u32, CAMIF_REQ_BUFS_MIN, rb->count); in s3c_camif_reqbufs() 924 ret = vb2_reqbufs(&vp->vb_queue, rb); in s3c_camif_reqbufs() 928 if (rb->count && rb->count < CAMIF_REQ_BUFS_MIN) { in s3c_camif_reqbufs() 929 rb->count = 0; in s3c_camif_reqbufs() 930 vb2_reqbufs(&vp->vb_queue, rb); in s3c_camif_reqbufs() 934 vp->reqbufs_count = rb->count; in s3c_camif_reqbufs() 935 if (vp->owner == NULL && rb->count > 0) in s3c_camif_reqbufs()
|
/linux-4.1.27/drivers/media/usb/stkwebcam/ |
D | stk-webcam.c | 1041 void *priv, struct v4l2_requestbuffers *rb) in stk_vidioc_reqbufs() argument 1047 if (rb->memory != V4L2_MEMORY_MMAP) in stk_vidioc_reqbufs() 1053 if (rb->count == 0) { in stk_vidioc_reqbufs() 1062 if (rb->count < 3) in stk_vidioc_reqbufs() 1063 rb->count = 3; in stk_vidioc_reqbufs() 1065 else if (rb->count > 5) in stk_vidioc_reqbufs() 1066 rb->count = 5; in stk_vidioc_reqbufs() 1068 stk_allocate_buffers(dev, rb->count); in stk_vidioc_reqbufs() 1069 rb->count = dev->n_sbufs; in stk_vidioc_reqbufs()
|
/linux-4.1.27/include/rdma/ |
D | ib_umem_odp.h | 42 struct rb_node rb; member
|
/linux-4.1.27/drivers/video/fbdev/omap2/dss/ |
D | manager-sysfs.c | 382 info.cpr_coefs.rb, in manager_cpr_coef_show() 403 &coefs.rr, &coefs.rg, &coefs.rb, in manager_cpr_coef_store() 408 arr = (s16[]){ coefs.rr, coefs.rg, coefs.rb, in manager_cpr_coef_store()
|
D | dispc.c | 1089 FLD_VAL(coefs->rb, 9, 0); in dispc_mgr_set_cpr_coef()
|
/linux-4.1.27/tools/perf/util/ |
D | kvm-stat.h | 25 struct rb_node rb; member
|
/linux-4.1.27/arch/arm/common/ |
D | dmabounce.c | 170 struct safe_buffer *b, *rb = NULL; in find_safe_buffer() local 178 rb = b; in find_safe_buffer() 183 return rb; in find_safe_buffer()
|
/linux-4.1.27/arch/x86/crypto/ |
D | aes-x86_64-asm_64.S | 83 #define round(TAB,OFFSET,r1,r2,r3,r4,r5,r6,r7,r8,ra,rb,rc,rd) \ argument 94 xorl OFFSET+4(r8),rb ## E; \
|
/linux-4.1.27/drivers/net/ethernet/ti/ |
D | netcp_ethss.c | 161 #define GBE_SET_REG_OFS(p, rb, rn) p->rb##_ofs.rn = \ argument 162 offsetof(struct gbe##_##rb, rn) 163 #define GBENU_SET_REG_OFS(p, rb, rn) p->rb##_ofs.rn = \ argument 164 offsetof(struct gbenu##_##rb, rn) 165 #define XGBE_SET_REG_OFS(p, rb, rn) p->rb##_ofs.rn = \ argument 166 offsetof(struct xgbe##_##rb, rn) 167 #define GBE_REG_ADDR(p, rb, rn) (p->rb + p->rb##_ofs.rn) argument
|
/linux-4.1.27/sound/pci/hda/ |
D | hda_controller.c | 989 PAGE_SIZE, &chip->rb); in azx_alloc_cmd_io() 998 chip->corb.addr = chip->rb.addr; in azx_init_cmd_io() 999 chip->corb.buf = (u32 *)chip->rb.area; in azx_init_cmd_io() 1035 chip->rirb.addr = chip->rb.addr + 2048; in azx_init_cmd_io() 1036 chip->rirb.buf = (u32 *)(chip->rb.area + 2048); in azx_init_cmd_io() 1508 if (chip->rb.area) in azx_free_stream_pages() 1509 chip->ops->dma_free_pages(chip, &chip->rb); in azx_free_stream_pages()
|
D | hda_controller.h | 341 struct snd_dma_buffer rb; member
|
/linux-4.1.27/include/drm/ |
D | drm_modes.h | 160 bool rb; member
|
D | drm_crtc.h | 1456 bool rb);
|
/linux-4.1.27/include/media/ |
D | v4l2-mem2mem.h | 240 struct v4l2_requestbuffers *rb);
|
/linux-4.1.27/block/ |
D | cfq-iosched.c | 85 struct rb_root rb; member 91 #define CFQ_RB_ROOT (struct cfq_rb_root) { .rb = RB_ROOT, \ 1171 root->left = rb_first(&root->rb); in cfq_rb_first() 1182 root->left = rb_first(&root->rb); in cfq_rb_first_group() 1200 rb_erase_init(n, &root->rb); in cfq_rb_erase() 1250 struct rb_node **node = &st->rb.rb_node; in __cfq_group_service_tree_add() 1272 rb_insert_color(&cfqg->rb_node, &st->rb); in __cfq_group_service_tree_add() 1365 n = rb_last(&st->rb); in cfq_group_notify_queue_add() 2035 parent = rb_last(&st->rb); in cfq_service_tree_add() 2072 p = &st->rb.rb_node; in cfq_service_tree_add() [all …]
|
/linux-4.1.27/drivers/media/platform/coda/ |
D | coda.h | 183 int (*reqbufs)(struct coda_ctx *ctx, struct v4l2_requestbuffers *rb);
|
D | coda-bit.c | 720 struct v4l2_requestbuffers *rb) in coda_encoder_reqbufs() argument 725 if (rb->type != V4L2_BUF_TYPE_VIDEO_OUTPUT) in coda_encoder_reqbufs() 728 if (rb->count) { in coda_encoder_reqbufs() 1402 struct v4l2_requestbuffers *rb) in coda_decoder_reqbufs() argument 1407 if (rb->type != V4L2_BUF_TYPE_VIDEO_OUTPUT) in coda_decoder_reqbufs() 1410 if (rb->count) { in coda_decoder_reqbufs()
|
D | coda-common.c | 699 struct v4l2_requestbuffers *rb) in coda_reqbufs() argument 704 ret = v4l2_m2m_reqbufs(file, ctx->fh.m2m_ctx, rb); in coda_reqbufs() 712 if (rb->type == V4L2_BUF_TYPE_VIDEO_OUTPUT && ctx->ops->reqbufs) in coda_reqbufs() 713 return ctx->ops->reqbufs(ctx, rb); in coda_reqbufs()
|
/linux-4.1.27/arch/sh/kernel/ |
D | disassemble.c | 304 int rb = 0; in print_sh_insn() local 372 rb = nibs[n] & 0x07; in print_sh_insn() 424 printk("r%d_bank", rb); in print_sh_insn()
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
D | radeon_mn.c | 79 it.rb) { in radeon_mn_destroy()
|
D | radeon_vm.c | 1234 rbtree_postorder_for_each_entry_safe(bo_va, tmp, &vm->va, it.rb) { in radeon_vm_fini()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
D | v4l2-mem2mem.c | 762 struct v4l2_requestbuffers *rb) in v4l2_m2m_ioctl_reqbufs() argument 766 return v4l2_m2m_reqbufs(file, fh->m2m_ctx, rb); in v4l2_m2m_ioctl_reqbufs()
|
/linux-4.1.27/drivers/media/pci/ttpci/ |
D | av7110_av.c | 443 #define FREE_COND_TS (dvb_ringbuffer_free(rb) >= 4096) 448 struct dvb_ringbuffer *rb; in ts_play() local 454 rb = (type) ? &av7110->avout : &av7110->aout; in ts_play() 467 if (wait_event_interruptible(rb->queue, FREE_COND_TS)) in ts_play()
|
/linux-4.1.27/drivers/media/usb/tm6000/ |
D | tm6000.h | 370 struct v4l2_requestbuffers *rb);
|
/linux-4.1.27/net/sched/ |
D | sch_htb.c | 350 static void htb_safe_rb_erase(struct rb_node *rb, struct rb_root *root) in htb_safe_rb_erase() argument 352 if (RB_EMPTY_NODE(rb)) { in htb_safe_rb_erase() 355 rb_erase(rb, root); in htb_safe_rb_erase() 356 RB_CLEAR_NODE(rb); in htb_safe_rb_erase()
|
D | sch_netem.c | 153 static struct sk_buff *netem_rb_to_skb(struct rb_node *rb) in netem_rb_to_skb() argument 155 return container_of(rb, struct sk_buff, rbnode); in netem_rb_to_skb()
|
/linux-4.1.27/drivers/scsi/ |
D | initio.c | 369 u8 instr, rb; in initio_se2_rd() local 382 rb = inb(base + TUL_NVRAM); in initio_se2_rd() 383 rb &= SE2DI; in initio_se2_rd() 384 val += (rb << i); in initio_se2_rd() 404 u8 rb; in initio_se2_wr() local 433 if ((rb = inb(base + TUL_NVRAM)) & SE2DI) in initio_se2_wr()
|
/linux-4.1.27/drivers/media/platform/ |
D | timblogiw.c | 285 struct v4l2_requestbuffers *rb) in timblogiw_reqbufs() argument 292 return videobuf_reqbufs(&fh->vb_vidq, rb); in timblogiw_reqbufs()
|
D | via-camera.c | 998 struct v4l2_requestbuffers *rb) in viacam_reqbufs() argument 1002 return videobuf_reqbufs(&cam->vb_queue, rb); in viacam_reqbufs()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | smc37c669.c | 945 #define rb( _x_ ) inb( (unsigned int)((unsigned long)_x_) ) macro 2013 data = rb( &SMC37c669->data_port ); in SMC37c669_read_config()
|
/linux-4.1.27/drivers/isdn/i4l/ |
D | isdn_tty.c | 2676 char rb[100]; in isdn_tty_cmd_ATand() local 2678 #define MAXRB (sizeof(rb) - 1) in isdn_tty_cmd_ATand() 2812 sprintf(rb, "S%02d=%03d%s", i, in isdn_tty_cmd_ATand() 2814 isdn_tty_at_cout(rb, info); in isdn_tty_cmd_ATand() 2816 sprintf(rb, "\r\nEAZ/MSN: %.50s\r\n", in isdn_tty_cmd_ATand() 2818 isdn_tty_at_cout(rb, info); in isdn_tty_cmd_ATand()
|
/linux-4.1.27/drivers/media/pci/cx18/ |
D | cx18-ioctl.c | 870 struct v4l2_requestbuffers *rb) in cx18_reqbufs() argument 880 return videobuf_reqbufs(cx18_vb_queue(id), rb); in cx18_reqbufs()
|
/linux-4.1.27/drivers/staging/media/omap4iss/ |
D | iss_video.c | 749 iss_video_reqbufs(struct file *file, void *fh, struct v4l2_requestbuffers *rb) in iss_video_reqbufs() argument 753 return vb2_reqbufs(&vfh->queue, rb); in iss_video_reqbufs()
|
/linux-4.1.27/drivers/xen/ |
D | xen-scsiback.c | 1586 ssize_t rb; in scsiback_tpg_param_show_alias() local 1589 rb = snprintf(page, PAGE_SIZE, "%s\n", tpg->param_alias); in scsiback_tpg_param_show_alias() 1592 return rb; in scsiback_tpg_param_show_alias()
|
/linux-4.1.27/drivers/media/usb/dvb-usb/ |
D | dib0700_devices.c | 2207 u8 rb[2]; in dib01x0_pmu_update() local 2210 {.addr = 0x1e >> 1, .flags = I2C_M_RD, .buf = rb, .len = 2}, in dib01x0_pmu_update() 2219 switch (rb[0] << 8 | rb[1]) { in dib01x0_pmu_update() 2244 wb[2] |= rb[0]; in dib01x0_pmu_update() 2245 wb[3] |= rb[1] & ~(3 << 4); in dib01x0_pmu_update()
|
/linux-4.1.27/drivers/media/platform/omap3isp/ |
D | ispvideo.c | 821 isp_video_reqbufs(struct file *file, void *fh, struct v4l2_requestbuffers *rb) in isp_video_reqbufs() argument 828 ret = vb2_reqbufs(&vfh->queue, rb); in isp_video_reqbufs()
|
/linux-4.1.27/fs/reiserfs/ |
D | fix_node.c | 624 int rnum, int blk_num, short *s012, int lb, int rb) in set_parameters() argument 641 tb->rbytes = rb; in set_parameters() 647 PROC_INFO_ADD(tb->tb_sb, rbytes[h], rb); in set_parameters()
|
/linux-4.1.27/arch/m68k/ifpsp060/src/ |
D | ilsp.S | 301 mov.w %d6, %d5 # rb + u3 307 mov.w %d6, %d5 # rb + u4
|
/linux-4.1.27/include/video/ |
D | omapdss.h | 380 s16 rr, rg, rb; member
|
/linux-4.1.27/drivers/media/usb/cx231xx/ |
D | cx231xx-video.c | 1609 struct v4l2_requestbuffers *rb) in vidioc_reqbufs() argument 1619 return videobuf_reqbufs(&fh->vb_vidq, rb); in vidioc_reqbufs()
|
/linux-4.1.27/Documentation/timers/ |
D | highres.txt | 38 - time ordered enqueueing into a rb-tree
|
/linux-4.1.27/Documentation/filesystems/ |
D | f2fs.txt | 143 extent_cache Enable an extent cache based on rb-tree, it can cache
|