/linux-4.4.14/lib/ |
H A D | list_debug.c | 19 * the prev/next entries already! 24 struct list_head *next) __list_add() 26 WARN(next->prev != prev, __list_add() 27 "list_add corruption. next->prev should be " __list_add() 28 "prev (%p), but was %p. (next=%p).\n", __list_add() 29 prev, next->prev, next); __list_add() 30 WARN(prev->next != next, __list_add() 31 "list_add corruption. prev->next should be " __list_add() 32 "next (%p), but was %p. (prev=%p).\n", __list_add() 33 next, prev->next, prev); __list_add() 34 WARN(new == prev || new == next, __list_add() 35 "list_add double add: new=%p, prev=%p, next=%p.\n", __list_add() 36 new, prev, next); __list_add() 37 next->prev = new; __list_add() 38 new->next = next; __list_add() 40 prev->next = new; __list_add() 46 struct list_head *prev, *next; __list_del_entry() local 49 next = entry->next; __list_del_entry() 51 if (WARN(next == LIST_POISON1, __list_del_entry() 52 "list_del corruption, %p->next is LIST_POISON1 (%p)\n", __list_del_entry() 57 WARN(prev->next != entry, __list_del_entry() 58 "list_del corruption. prev->next should be %p, " __list_del_entry() 59 "but was %p\n", entry, prev->next) || __list_del_entry() 60 WARN(next->prev != entry, __list_del_entry() 61 "list_del corruption. next->prev should be %p, " __list_del_entry() 62 "but was %p\n", entry, next->prev)) __list_del_entry() 65 __list_del(prev, next); __list_del_entry() 78 entry->next = LIST_POISON1; list_del() 87 struct list_head *prev, struct list_head *next) __list_add_rcu() 89 WARN(next->prev != prev, __list_add_rcu() 90 "list_add_rcu corruption. next->prev should be prev (%p), but was %p. (next=%p).\n", __list_add_rcu() 91 prev, next->prev, next); __list_add_rcu() 92 WARN(prev->next != next, __list_add_rcu() 93 "list_add_rcu corruption. prev->next should be next (%p), but was %p. (prev=%p).\n", __list_add_rcu() 94 next, prev->next, prev); __list_add_rcu() 95 new->next = next; __list_add_rcu() 98 next->prev = new; __list_add_rcu() 22 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 86 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
|
H A D | plist.c | 36 WARN(n->prev != p || p->next != n, plist_check_prev_next() 39 "next: %p, n: %p, p: %p\n", plist_check_prev_next() 40 t, t->next, t->prev, plist_check_prev_next() 41 p, p->next, p->prev, plist_check_prev_next() 42 n, n->next, n->prev); plist_check_prev_next() 47 struct list_head *prev = top, *next = top->next; plist_check_list() local 49 plist_check_prev_next(top, prev, next); plist_check_list() 50 while (next != top) { plist_check_list() 51 prev = next; plist_check_list() 52 next = prev->next; plist_check_list() 53 plist_check_prev_next(top, prev, next); plist_check_list() 95 iter = list_entry(iter->prio_list.next, plist_add() 118 if (node->node_list.next != &head->node_list) { plist_del() 119 struct plist_node *next; plist_del() local 121 next = list_entry(node->node_list.next, plist_del() 124 /* add the next plist_node into prio_list */ plist_del() 125 if (list_empty(&next->prio_list)) plist_del() 126 list_add(&next->prio_list, &node->prio_list); plist_del() 204 BUG_ON(prio_pos->prio_list.next != &node_pos->prio_list); plist_test_check() 209 BUG_ON(prio_pos->prio_list.next != &first->prio_list); plist_test_check()
|
H A D | ioremap.c | 75 unsigned long next; ioremap_pmd_range() local 82 next = pmd_addr_end(addr, end); ioremap_pmd_range() 85 ((next - addr) == PMD_SIZE) && ioremap_pmd_range() 91 if (ioremap_pte_range(pmd, addr, next, phys_addr + addr, prot)) ioremap_pmd_range() 93 } while (pmd++, addr = next, addr != end); ioremap_pmd_range() 101 unsigned long next; ioremap_pud_range() local 108 next = pud_addr_end(addr, end); ioremap_pud_range() 111 ((next - addr) == PUD_SIZE) && ioremap_pud_range() 117 if (ioremap_pmd_range(pud, addr, next, phys_addr + addr, prot)) ioremap_pud_range() 119 } while (pud++, addr = next, addr != end); ioremap_pud_range() 128 unsigned long next; ioremap_page_range() local 137 next = pgd_addr_end(addr, end); ioremap_page_range() 138 err = ioremap_pud_range(pgd, addr, next, phys_addr+addr, prot); ioremap_page_range() 141 } while (pgd++, addr = next, addr != end); ioremap_page_range()
|
H A D | list_sort.c | 29 tail->next = a; merge() 30 a = a->next; merge() 32 tail->next = b; merge() 33 b = b->next; merge() 35 tail = tail->next; merge() 37 tail->next = a?:b; merge() 38 return head.next; merge() 60 tail->next = a; merge_and_restore_back_links() 62 a = a->next; merge_and_restore_back_links() 64 tail->next = b; merge_and_restore_back_links() 66 b = b->next; merge_and_restore_back_links() 68 tail = tail->next; merge_and_restore_back_links() 70 tail->next = a ? : b; merge_and_restore_back_links() 80 (*cmp)(priv, tail->next, tail->next); merge_and_restore_back_links() 82 tail->next->prev = tail; merge_and_restore_back_links() 83 tail = tail->next; merge_and_restore_back_links() 84 } while (tail->next); merge_and_restore_back_links() 86 tail->next = head; merge_and_restore_back_links() 119 head->prev->next = NULL; list_sort() 120 list = head->next; list_sort() 124 list = list->next; list_sort() 125 cur->next = NULL; list_sort() 245 for (cur = head.next; cur->next != &head; cur = cur->next) { list_sort_test() 249 if (cur->next->prev != cur) { list_sort_test() 254 cmp_result = cmp(NULL, cur, cur->next); list_sort_test() 261 el1 = container_of(cur->next, struct debug_el, list); list_sort_test()
|
H A D | timerqueue.c | 59 if (!head->next || node->expires.tv64 < head->next->expires.tv64) { timerqueue_add() 60 head->next = node; timerqueue_add() 79 /* update next pointer */ timerqueue_del() 80 if (head->next == node) { timerqueue_del() 83 head->next = rbn ? timerqueue_del() 88 return head->next != NULL; timerqueue_del() 103 struct rb_node *next; timerqueue_iterate_next() local 107 next = rb_next(&node->node); timerqueue_iterate_next() 108 if (!next) timerqueue_iterate_next() 110 return container_of(next, struct timerqueue_node, node); timerqueue_iterate_next()
|
H A D | llist.c | 44 new_last->next = first = ACCESS_ONCE(head->first); llist_add_batch() 61 * llist_add) sequence in another user may change @head->first->next, 67 struct llist_node *entry, *old_entry, *next; llist_del_first() local 74 next = READ_ONCE(entry->next); llist_del_first() 75 entry = cmpxchg(&head->first, old_entry, next); llist_del_first() 97 head = head->next; llist_reverse_order() 98 tmp->next = new_head; llist_reverse_order()
|
H A D | klist.c | 23 * next element in the list. It returns NULL if there are no more items. 26 * the next klist_node. It then drops the lock and returns. 368 * klist_next - Ante up next node in list. 372 * node, if there was one. Grab the next node, increment its reference 373 * count, drop the lock, and return that next node. 379 struct klist_node *next; klist_next() local 384 next = to_klist_node(last->n_node.next); klist_next() 388 next = to_klist_node(i->i_klist->k_list.next); klist_next() 391 while (next != to_klist_node(&i->i_klist->k_list)) { klist_next() 392 if (likely(!knode_dead(next))) { klist_next() 393 kref_get(&next->n_ref); klist_next() 394 i->i_cur = next; klist_next() 397 next = to_klist_node(next->n_node.next); klist_next()
|
/linux-4.4.14/arch/score/include/asm/ |
H A D | switch_to.h | 4 extern void *resume(void *last, void *next, void *next_ti); 6 #define switch_to(prev, next, last) \ 8 (last) = resume(prev, next, task_thread_info(next)); \
|
H A D | mmu_context.h | 72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 78 if ((next->context ^ asid_cache) & ASID_VERSION_MASK) switch_mm() 79 get_new_mmu_context(next); switch_mm() 81 pevn_set(next->context); switch_mm() 82 TLBMISS_HANDLER_SETUP_PGD(next->pgd); switch_mm() 102 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 107 get_new_mmu_context(next); activate_mm() 108 pevn_set(next->context); activate_mm() 109 TLBMISS_HANDLER_SETUP_PGD(next->pgd); activate_mm()
|
/linux-4.4.14/tools/firewire/ |
H A D | list.h | 2 struct list *next, *prev; member in struct:list 8 list->next = list; list_init() 15 return list->next == list; list_empty() 22 new_link->next = link; list_insert() 23 new_link->prev->next = new_link; list_insert() 24 new_link->next->prev = new_link; list_insert() 36 list_insert(list->next, new_link); list_prepend() 42 link->prev->next = link->next; list_remove() 43 link->next->prev = link->prev; list_remove() 50 list_entry((list)->next, type, member) 56 list_entry((elm)->member.next, typeof(*elm), member)
|
/linux-4.4.14/arch/parisc/include/asm/ |
H A D | switch_to.h | 8 #define switch_to(prev, next, last) do { \ 9 (last) = _switch_to(prev, next); \
|
H A D | mmu_context.h | 52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument 55 if (prev != next) { switch_mm() 56 mtctl(__pa(next->pgd), 25); switch_mm() 57 load_context(next->context); switch_mm() 63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 73 BUG_ON(next == &init_mm); /* Should never happen */ activate_mm() 75 if (next->context == 0) activate_mm() 76 next->context = alloc_sid(); activate_mm() 78 switch_mm(prev,next,current); activate_mm()
|
/linux-4.4.14/kernel/locking/ |
H A D | osq_lock.c | 32 * Get a stable @node->next pointer, either for unlock() or unqueue() purposes. 40 struct optimistic_spin_node *next = NULL; osq_wait_next() local 63 * We must xchg() the @node->next value, because if we were to osq_wait_next() 65 * @node->next might complete Step-A and think its @prev is osq_wait_next() 70 * wait for a new @node->next from its Step-C. osq_wait_next() 72 if (node->next) { osq_wait_next() 73 next = xchg(&node->next, NULL); osq_wait_next() 74 if (next) osq_wait_next() 81 return next; osq_wait_next() 87 struct optimistic_spin_node *prev, *next; osq_lock() local 92 node->next = NULL; osq_lock() 107 WRITE_ONCE(prev->next, node); osq_lock() 133 * Undo our @prev->next assignment; this will make @prev's osq_lock() 134 * unlock()/unqueue() wait for a next pointer since @lock points to us osq_lock() 139 if (prev->next == node && osq_lock() 140 cmpxchg(&prev->next, node, NULL) == node) osq_lock() 161 * Step - B -- stabilize @next osq_lock() 163 * Similar to unlock(), wait for @node->next or move @lock from @node osq_lock() 167 next = osq_wait_next(lock, node, prev); osq_lock() 168 if (!next) osq_lock() 174 * @prev is stable because its still waiting for a new @prev->next osq_lock() 175 * pointer, @next is stable because our @node->next pointer is NULL and osq_lock() 179 WRITE_ONCE(next->prev, prev); osq_lock() 180 WRITE_ONCE(prev->next, next); osq_lock() 187 struct optimistic_spin_node *node, *next; osq_unlock() local 201 next = xchg(&node->next, NULL); osq_unlock() 202 if (next) { osq_unlock() 203 WRITE_ONCE(next->locked, 1); osq_unlock() 207 next = osq_wait_next(lock, node, NULL); osq_unlock() 208 if (next) osq_unlock() 209 WRITE_ONCE(next->locked, 1); osq_unlock()
|
H A D | mcs_spinlock.h | 18 struct mcs_spinlock *next; member in struct:mcs_spinlock 68 node->next = NULL; mcs_spin_lock() 88 WRITE_ONCE(prev->next, node); mcs_spin_lock() 101 struct mcs_spinlock *next = READ_ONCE(node->next); mcs_spin_unlock() local 103 if (likely(!next)) { mcs_spin_unlock() 109 /* Wait until the next pointer is set */ mcs_spin_unlock() 110 while (!(next = READ_ONCE(node->next))) mcs_spin_unlock() 114 /* Pass lock to next waiter. */ mcs_spin_unlock() 115 arch_mcs_spin_unlock_contended(&next->locked); mcs_spin_unlock()
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | switch_to.h | 31 #define switch_to(prev,next,last) do { \ 37 if (next->mm) { \ 38 update_cr_regs(next); \ 40 restore_access_regs(&next->thread.acrs[0]); \ 41 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \ 43 prev = __switch_to(prev,next); \
|
H A D | mmu_context.h | 85 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 90 S390_lowcore.user_asce = next->context.asce; switch_mm() 91 if (prev == next) switch_mm() 94 cpumask_set_cpu(cpu, &next->context.cpu_attach_mask); switch_mm() 98 atomic_inc(&next->context.attach_count); switch_mm() 128 struct mm_struct *next) activate_mm() 130 switch_mm(prev, next, current); activate_mm() 131 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); activate_mm() 132 set_user_asce(next); activate_mm() 127 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | switch_to_64.h | 21 struct task_struct *next, 24 #define switch_to(prev,next,last) \ 26 if (last_task_used_math != next) { \ 27 struct pt_regs *regs = next->thread.uregs; \ 30 last = sh64_switch_to(prev, &prev->thread, next, \ 31 &next->thread); \
|
H A D | switch_to_32.h | 64 struct task_struct *next); 69 #define switch_to(prev, next, last) \ 81 if (is_dsp_enabled(next)) \ 82 __restore_dsp(next); \ 87 __ts5 = (u32 *)next; \ 88 __ts6 = (u32 *)&next->thread.sp; \ 89 __ts7 = next->thread.pc; \
|
H A D | mmu_context.h | 119 struct mm_struct *next, switch_mm() 124 if (likely(prev != next)) { switch_mm() 125 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm() 126 set_TTB(next->pgd); switch_mm() 127 activate_context(next, cpu); switch_mm() 129 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) switch_mm() 130 activate_context(next, cpu); switch_mm() 133 #define activate_mm(prev, next) switch_mm((prev),(next),NULL) 118 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
|
/linux-4.4.14/arch/blackfin/include/asm/ |
H A D | switch_to.h | 21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next); 24 #define switch_to(prev,next,last) \ 28 memcpy (L1_SCRATCH_TASK_INFO, &task_thread_info(next)->l1_task_info, \ 30 (last) = resume (prev, next); \ 33 #define switch_to(prev, next, last) \ 35 (last) = resume(prev, next); \
|
H A D | mmu.h | 12 struct sram_list_struct *next; member in struct:sram_list_struct
|
/linux-4.4.14/arch/cris/include/asm/ |
H A D | switch_to.h | 8 extern struct task_struct *resume(struct task_struct *prev, struct task_struct *next, int); 9 #define switch_to(prev,next,last) last = resume(prev,next, \
|
H A D | mmu_context.h | 9 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 14 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 19 switch_mm(prev, next, NULL); activate_mm()
|
/linux-4.4.14/tools/perf/ui/ |
H A D | progress.c | 19 if (p->curr >= p->next) { ui_progress__update() 20 p->next += p->step; ui_progress__update() 28 p->next = p->step = total / 16; ui_progress__init()
|
H A D | progress.h | 10 u64 curr, next, step, total; member in struct:ui_progress
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | switch_to.h | 17 * switch_to(prev, next) should switch from task `prev' to `next' 18 * `prev' will never be the same as `next'. schedule() itself 24 #define switch_to(prev, next, last) \ 27 task_thread_info(next)); \
|
H A D | mmu_context.h | 49 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 54 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) switch_mm() 55 cpu_switch_mm(next->pgd, next); switch_mm() 59 #define activate_mm(prev, next) switch_mm(prev, next, NULL)
|
/linux-4.4.14/sound/pci/ctxfi/ |
H A D | ctimap.c | 31 entry->next = entry->addr; input_mapper_add() 53 pos = head->next; 60 entry->next = pos_ent->addr; 62 pre_ent->next = entry->addr; 71 struct list_head *next, *pre, *head; input_mapper_delete() local 80 next = (entry->list.next == head) ? head->next : entry->list.next; input_mapper_delete() 84 entry->next = entry->addr = entry->user = entry->slot = 0; input_mapper_delete() 91 next_ent = list_entry(next, struct imapper, list); input_mapper_delete() 93 pre_ent->next = next_ent->addr; input_mapper_delete() 106 pos = head->next; free_input_mapper_list()
|
H A D | ctimap.h | 28 unsigned short next; /* the next input mapper ram id */ member in struct:imapper
|
/linux-4.4.14/ipc/ |
H A D | msgutil.c | 27 * The next 2 defines are here bc this is the only file 43 struct msg_msgseg *next; member in struct:msg_msgseg 44 /* the next part of the message follows immediately */ 62 msg->next = NULL; alloc_msg() 66 pseg = &msg->next; alloc_msg() 74 seg->next = NULL; alloc_msg() 75 pseg = &seg->next; alloc_msg() 101 for (seg = msg->next; seg != NULL; seg = seg->next) { load_msg() 132 for (dst_pseg = dst->next, src_pseg = src->next; copy_msg() 134 dst_pseg = dst_pseg->next, src_pseg = src_pseg->next) { copy_msg() 161 for (seg = msg->next; seg != NULL; seg = seg->next) { store_msg() 177 seg = msg->next; free_msg() 180 struct msg_msgseg *tmp = seg->next; free_msg()
|
/linux-4.4.14/tools/usb/usbip/libsrc/ |
H A D | list.h | 13 * sometimes we already know the next/prev entries and we can 19 struct list_head *next, *prev; member in struct:list_head 29 list->next = list; INIT_LIST_HEAD() 37 * the prev/next entries already! 41 struct list_head *next) __list_add() 43 next->prev = new; __list_add() 44 new->next = next; __list_add() 46 prev->next = new; __list_add() 59 __list_add(new, head, head->next); list_add() 63 * Delete a list entry by making the prev/next entries 67 * the prev/next entries already! 69 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument 71 next->prev = prev; __list_del() 72 prev->next = next; __list_del() 87 __list_del(entry->prev, entry->next); __list_del_entry() 92 __list_del(entry->prev, entry->next); list_del() 93 entry->next = LIST_POISON1; list_del() 111 for (pos = (head)->next; pos != (head); pos = pos->next) 120 for (pos = (head)->next, n = pos->next; pos != (head); \ 121 pos = n, n = pos->next) 39 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
|
H A D | names.c | 44 struct vendor *next; member in struct:vendor 50 struct product *next; member in struct:product 56 struct class *next; member in struct:class 62 struct subclass *next; member in struct:subclass 68 struct protocol *next; member in struct:protocol 74 struct genericstrtable *next; member in struct:genericstrtable 106 for (; v; v = v->next) names_vendor() 117 for (; p; p = p->next) names_product() 128 for (; c; c = c->next) names_class() 139 for (; s; s = s->next) names_subclass() 152 for (; p; p = p->next) names_protocol() 161 struct pool *next; member in struct:pool 181 p->next = pool_head; my_malloc() 201 pool = pool->next; names_free() 212 for (; v; v = v->next) new_vendor() 220 v->next = vendors[h]; new_vendor() 232 for (; p; p = p->next) new_product() 241 p->next = products[h]; new_product() 252 for (; c; c = c->next) new_class() 260 c->next = classes[h]; new_class() 271 for (; s; s = s->next) new_subclass() 280 s->next = subclasses[h]; new_subclass() 293 for (; p; p = p->next) new_protocol() 304 p->next = protocols[h]; new_protocol()
|
/linux-4.4.14/drivers/scsi/sym53c8xx_2/ |
H A D | sym_malloc.c | 77 while (!h[j].next) { ___sym_malloc() 79 h[j].next = (m_link_p) M_GET_MEM_CLUSTER(); ___sym_malloc() 80 if (h[j].next) ___sym_malloc() 81 h[j].next->next = NULL; ___sym_malloc() 87 a = h[j].next; ___sym_malloc() 89 h[j].next = h[j].next->next; ___sym_malloc() 93 h[j].next = (m_link_p) (a+s); ___sym_malloc() 94 h[j].next->next = NULL; ___sym_malloc() 133 ((m_link_p) a)->next = h[i].next; ___sym_mfree() 134 h[i].next = (m_link_p) a; ___sym_mfree() 140 while (q->next && q->next != (m_link_p) b) { ___sym_mfree() 141 q = q->next; ___sym_mfree() 143 if (!q->next) { ___sym_mfree() 144 ((m_link_p) a)->next = h[i].next; ___sym_mfree() 145 h[i].next = (m_link_p) a; ___sym_mfree() 148 q->next = q->next->next; ___sym_mfree() 235 vbp->next = mp->vtob[hc]; ___get_dma_mem_cluster() 253 vbpp = &(*vbpp)->next; ___free_dma_mem_cluster() 256 *vbpp = (*vbpp)->next; ___free_dma_mem_cluster() 268 for (mp = mp0.next; ___get_dma_pool() 270 mp = mp->next); ___get_dma_pool() 284 mp->next = mp0.next; ___cre_dma_pool() 285 mp0.next = mp; ___cre_dma_pool() 295 m_pool_p *pp = &mp0.next; ___del_dma_pool() 298 pp = &(*pp)->next; ___del_dma_pool() 300 *pp = (*pp)->next; ___del_dma_pool() 371 vp = vp->next; __vtobus()
|
/linux-4.4.14/drivers/thunderbolt/ |
H A D | cap.c | 30 && cap->extended_short.next == 0 tb_cap_is_long() 45 int next; tb_cap_next() local 51 next = cap->basic.next; tb_cap_next() 59 next = cap->basic.next; tb_cap_next() 61 next = cap->extended_short.next; tb_cap_next() 63 next = cap->extended_long.next; tb_cap_next() 69 if (next == offset) tb_cap_next() 71 return next; tb_cap_next()
|
/linux-4.4.14/arch/xtensa/include/asm/ |
H A D | switch_to.h | 15 extern void *_switch_to(void *last, void *next); 17 #define switch_to(prev,next,last) \ 19 (last) = _switch_to(prev, next); \
|
H A D | nommu_context.h | 18 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 22 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
|
H A D | mmu_context.h | 73 * Start new asid cycle; continue counting with next get_new_mmu_context() 123 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 127 int migrated = next->context.cpu != cpu; switch_mm() 131 next->context.cpu = cpu; switch_mm() 133 if (migrated || prev != next) switch_mm() 134 activate_context(next, cpu); switch_mm() 137 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.4.14/arch/microblaze/include/asm/ |
H A D | switch_to.h | 16 struct thread_info *next); 18 #define switch_to(prev, next, last) \ 21 task_thread_info(next)); \
|
/linux-4.4.14/drivers/pnp/isapnp/ |
H A D | compat.c | 33 list = from ? from->global_list.next : pnp_cards.next; pnp_find_card() 40 list = list->next; pnp_find_card() 56 list = pnp_global.next; pnp_find_dev() 58 list = from->global_list.next; pnp_find_dev() 66 list = list->next; pnp_find_dev() 71 list = card->devices.next; pnp_find_dev() 73 list = from->card_list.next; pnp_find_dev() 82 list = list->next; pnp_find_dev()
|
/linux-4.4.14/arch/arc/include/asm/ |
H A D | switch_to.h | 31 #define switch_to(prev, next, last) \ 33 ARC_FPU_PREV(prev, next); \ 34 last = __switch_to(prev, next);\ 35 ARC_FPU_NEXT(next); \
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | switch_to.h | 18 #define ocd_switch(prev, next) \ 21 ocd_write(PID, next->pid); \ 24 #define ocd_switch(prev, next) 28 * switch_to(prev, next, last) should switch from task `prev' to task 29 * `next'. `prev' will never be the same as `next'. 41 #define switch_to(prev, next, last) \ 43 ocd_switch(prev, next); \ 45 &next->thread.cpu_context); \
|
H A D | mmu_context.h | 109 struct mm_struct *next, switch_mm() 112 if (likely(prev != next)) { switch_mm() 113 unsigned long __pgdir = (unsigned long)next->pgd; switch_mm() 116 activate_context(next); switch_mm() 122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL) 108 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
|
/linux-4.4.14/tools/power/cpupower/lib/ |
H A D | cpufreq.c | 80 struct cpufreq_available_governors *tmp, *next; cpufreq_put_available_governors() local 87 next = tmp->next; cpufreq_put_available_governors() 91 tmp = next; cpufreq_put_available_governors() 104 struct cpufreq_available_frequencies *tmp, *next; cpufreq_put_available_frequencies() local 111 next = tmp->next; cpufreq_put_available_frequencies() 113 tmp = next; cpufreq_put_available_frequencies() 125 struct cpufreq_affected_cpus *tmp, *next; cpufreq_put_affected_cpus() local 132 next = tmp->next; cpufreq_put_affected_cpus() 134 tmp = next; cpufreq_put_affected_cpus() 192 struct cpufreq_stats *tmp, *next; cpufreq_put_stats() local 199 next = tmp->next; cpufreq_put_stats() 201 tmp = next; cpufreq_put_stats()
|
/linux-4.4.14/block/ |
H A D | cmdline-parser.c | 42 char *next = strchr(++partdef, ')'); parse_subpart() local 44 if (!next) { parse_subpart() 50 length = min_t(int, next - partdef, parse_subpart() 55 partdef = ++next; parse_subpart() 92 char *next; parse_parts() local 104 next = strchr(bdevdef, ':'); parse_parts() 105 if (!next) { parse_parts() 110 length = min_t(int, next - bdevdef, sizeof(newparts->name) - 1); parse_parts() 117 while (next && *(++next)) { parse_parts() 118 bdevdef = next; parse_parts() 119 next = strchr(bdevdef, ','); parse_parts() 121 length = (!next) ? (sizeof(buf) - 1) : parse_parts() 122 min_t(int, next - bdevdef, sizeof(buf) - 1); parse_parts() 168 char *next; cmdline_parts_parse() local 173 next = pbuf = buf = kstrdup(cmdline, GFP_KERNEL); cmdline_parts_parse() 179 while (next && *pbuf) { cmdline_parts_parse() 180 next = strchr(pbuf, ';'); cmdline_parts_parse() 181 if (next) cmdline_parts_parse() 182 *next = '\0'; cmdline_parts_parse() 188 if (next) cmdline_parts_parse() 189 pbuf = ++next; cmdline_parts_parse()
|
H A D | blk-merge.c | 41 * If the next starting sector would be misaligned, stop the discard at blk_bio_discard_split() 554 struct request *next) ll_merge_requests_fn() 558 req->biotail->bi_seg_back_size + next->bio->bi_seg_front_size; ll_merge_requests_fn() 564 if (req_no_special_merge(req) || req_no_special_merge(next)) ll_merge_requests_fn() 567 if (req_gap_back_merge(req, next->bio)) ll_merge_requests_fn() 573 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > ll_merge_requests_fn() 577 total_phys_segments = req->nr_phys_segments + next->nr_phys_segments; ll_merge_requests_fn() 578 if (blk_phys_contig_segment(q, req->biotail, next->bio)) { ll_merge_requests_fn() 581 if (next->nr_phys_segments == 1) ll_merge_requests_fn() 582 next->biotail->bi_seg_back_size = seg_size; ll_merge_requests_fn() 589 if (blk_integrity_merge_rq(q, req, next) == false) ll_merge_requests_fn() 648 struct request *next) attempt_merge() 650 if (!rq_mergeable(req) || !rq_mergeable(next)) attempt_merge() 653 if (!blk_check_merge_flags(req->cmd_flags, next->cmd_flags)) attempt_merge() 659 if (blk_rq_pos(req) + blk_rq_sectors(req) != blk_rq_pos(next)) attempt_merge() 662 if (rq_data_dir(req) != rq_data_dir(next) attempt_merge() 663 || req->rq_disk != next->rq_disk attempt_merge() 664 || req_no_special_merge(next)) attempt_merge() 668 !blk_write_same_mergeable(req->bio, next->bio)) attempt_merge() 673 * from next to rq and release next. merge_requests_fn attempt_merge() 677 if (!ll_merge_requests_fn(q, req, next)) attempt_merge() 686 if ((req->cmd_flags | next->cmd_flags) & REQ_MIXED_MERGE || attempt_merge() 688 (next->cmd_flags & REQ_FAILFAST_MASK)) { attempt_merge() 690 blk_rq_set_mixed_merge(next); attempt_merge() 699 if (time_after(req->start_time, next->start_time)) attempt_merge() 700 req->start_time = next->start_time; attempt_merge() 702 req->biotail->bi_next = next->bio; attempt_merge() 703 req->biotail = next->biotail; attempt_merge() 705 req->__data_len += blk_rq_bytes(next); attempt_merge() 707 elv_merge_requests(q, req, next); attempt_merge() 710 * 'next' is going away, so update stats accordingly attempt_merge() 712 blk_account_io_merge(next); attempt_merge() 714 req->ioprio = ioprio_best(req->ioprio, next->ioprio); attempt_merge() 715 if (blk_rq_cpu_valid(next)) attempt_merge() 716 req->cpu = next->cpu; attempt_merge() 718 /* owner-ship of bio passed from next to req */ attempt_merge() 719 next->bio = NULL; attempt_merge() 720 __blk_put_request(q, next); attempt_merge() 726 struct request *next = elv_latter_request(q, rq); attempt_back_merge() local 728 if (next) attempt_back_merge() 729 return attempt_merge(q, rq, next); attempt_back_merge() 745 struct request *next) blk_attempt_req_merge() 747 return attempt_merge(q, rq, next); blk_attempt_req_merge() 553 ll_merge_requests_fn(struct request_queue *q, struct request *req, struct request *next) ll_merge_requests_fn() argument 647 attempt_merge(struct request_queue *q, struct request *req, struct request *next) attempt_merge() argument 744 blk_attempt_req_merge(struct request_queue *q, struct request *rq, struct request *next) blk_attempt_req_merge() argument
|
H A D | noop-iosched.c | 16 struct request *next) noop_merged_requests() 18 list_del_init(&next->queuelist); noop_merged_requests() 57 if (rq->queuelist.next == &nd->queue) noop_latter_request() 15 noop_merged_requests(struct request_queue *q, struct request *rq, struct request *next) noop_merged_requests() argument
|
/linux-4.4.14/net/ax25/ |
H A D | ax25_iface.c | 38 struct listen_struct *next; member in struct:listen_struct 51 ap->next = protocol_list; ax25_register_pid() 68 protocol_list = protocol->next; ax25_protocol_release() 72 while (protocol != NULL && protocol->next != NULL) { ax25_protocol_release() 73 if (protocol->next->pid == pid) { ax25_protocol_release() 74 protocol->next = protocol->next->next; ax25_protocol_release() 78 protocol = protocol->next; ax25_protocol_release() 118 listen->next = listen_list; ax25_listen_register() 139 listen_list = listen->next; ax25_listen_release() 145 while (listen != NULL && listen->next != NULL) { ax25_listen_release() 146 if (ax25cmp(&listen->next->callsign, callsign) == 0 && listen->next->dev == dev) { ax25_listen_release() 147 s = listen->next; ax25_listen_release() 148 listen->next = listen->next->next; ax25_listen_release() 154 listen = listen->next; ax25_listen_release() 167 for (protocol = protocol_list; protocol != NULL; protocol = protocol->next) ax25_protocol_function() 182 for (listen = listen_list; listen != NULL; listen = listen->next) ax25_listen_mine() 209 for (protocol = protocol_list; protocol != NULL; protocol = protocol->next) ax25_protocol_is_registered()
|
H A D | ax25_dev.c | 40 for (ax25_dev = ax25_dev_list; ax25_dev != NULL; ax25_dev = ax25_dev->next) ax25_addr_ax25dev() 87 ax25_dev->next = ax25_dev_list; ax25_dev_device_up() 112 for (s = ax25_dev_list; s != NULL; s = s->next) ax25_dev_device_down() 117 ax25_dev_list = s->next; ax25_dev_device_down() 124 while (s != NULL && s->next != NULL) { ax25_dev_device_down() 125 if (s->next == ax25_dev) { ax25_dev_device_down() 126 s->next = ax25_dev->next; ax25_dev_device_down() 133 s = s->next; ax25_dev_device_down() 193 ax25_dev = ax25_dev->next; ax25_dev_free()
|
H A D | ax25_route.c | 53 ax25_rt = ax25_rt->next; ax25_rt_device_down() 57 ax25_route_list = s->next; ax25_rt_device_down() 61 for (t = ax25_route_list; t != NULL; t = t->next) { ax25_rt_device_down() 62 if (t->next == s) { ax25_rt_device_down() 63 t->next = s->next; ax25_rt_device_down() 109 ax25_rt = ax25_rt->next; ax25_rt_add() 135 ax25_rt->next = ax25_route_list; ax25_rt_add() 161 ax25_rt = ax25_rt->next; ax25_rt_del() 165 ax25_route_list = s->next; ax25_rt_del() 168 for (t = ax25_route_list; t != NULL; t = t->next) { ax25_rt_del() 169 if (t->next == s) { ax25_rt_del() 170 t->next = s->next; ax25_rt_del() 216 ax25_rt = ax25_rt->next; ax25_rt_opt() 262 for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) { __acquires() 275 ((struct ax25_route *) v)->next; ax25_rt_seq_next() 328 .next = ax25_rt_seq_next, 364 for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) { ax25_get_route() 499 ax25_rt = ax25_rt->next; ax25_rt_free()
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | switch_to.h | 44 #define __switch_to(prev,next,last) do { \ 47 if (IA64_HAS_EXTRA_STATE(next)) \ 48 ia64_load_extra(next); \ 49 ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \ 50 (last) = ia64_switch_to((next)); \ 60 # define switch_to(prev,next,last) do { \ 66 __switch_to(prev, next, last); \ 67 /* "next" in old context is "current" in new context */ \ 76 # define switch_to(prev,next,last) __switch_to(prev, next, last)
|
H A D | mmu_context.h | 36 unsigned int next; /* next context number to use */ member in struct:ia64_ctx 39 /* call wrap_mmu_context when next >= max */ 91 if (ia64_ctx.next >= ia64_ctx.limit) { get_mmu_context() 92 ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, get_mmu_context() 93 ia64_ctx.max_ctx, ia64_ctx.next); get_mmu_context() 95 ia64_ctx.max_ctx, ia64_ctx.next); get_mmu_context() 96 if (ia64_ctx.next >= ia64_ctx.max_ctx) get_mmu_context() 99 mm->context = context = ia64_ctx.next++; get_mmu_context() 185 activate_mm (struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 191 ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd)); activate_mm() 192 activate_context(next); activate_mm()
|
/linux-4.4.14/scripts/kconfig/ |
H A D | list.h | 24 struct list_head *next, *prev; member in struct:list_head 49 for (pos = list_entry((head)->next, typeof(*pos), member); \ 51 pos = list_entry(pos->member.next, typeof(*pos), member)) 61 for (pos = list_entry((head)->next, typeof(*pos), member), \ 62 n = list_entry(pos->member.next, typeof(*pos), member); \ 64 pos = n, n = list_entry(n->member.next, typeof(*n), member)) 72 return head->next == head; list_empty() 79 * the prev/next entries already! 83 struct list_head *next) __list_add() 85 next->prev = _new; __list_add() 86 _new->next = next; __list_add() 88 prev->next = _new; __list_add() 105 * Delete a list entry by making the prev/next entries 109 * the prev/next entries already! 111 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 113 next->prev = prev; __list_del() 114 prev->next = next; __list_del() 127 __list_del(entry->prev, entry->next); list_del() 128 entry->next = (struct list_head*)LIST_POISON1; list_del() 81 __list_add(struct list_head *_new, struct list_head *prev, struct list_head *next) __list_add() argument
|
H A D | kxgettext.c | 43 goto next; escape() 50 next: escape() 64 struct file_line *next; member in struct:file_line 78 self->next = NULL; file_line__new() 86 struct message *next; member in struct:message 109 self->next = NULL; message__new() 127 m = m->next; mesage__find() 142 fl->next = self->files; message__add_file_line() 163 m->next = message__list; message__add() 184 for (child = menu->list; child != NULL; child = child->next) menu_build_message_list() 198 fl = fl->next; message__print_file_lineno() 202 fl = fl->next; message__print_file_lineno() 224 m = m->next; menu__xgettext()
|
/linux-4.4.14/arch/um/include/shared/ |
H A D | aio.h | 18 struct aio_context *next; member in struct:aio_context 22 .next = NULL }
|
H A D | irq_user.h | 12 struct irq_fd *next; member in struct:irq_fd
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | spinlock_types.h | 15 u16 next; member in struct:__anon191::__anon192::__raw_tickets 19 u16 next;
|
H A D | switch_to.h | 19 * switch_to(prev, next) should switch from task `prev' to `next' 20 * `prev' will never be the same as `next'. schedule() itself 25 #define switch_to(prev,next,last) \ 28 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
|
H A D | mmu_context.h | 93 #define activate_mm(prev,next) switch_mm(prev, next, NULL) 116 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 128 !cpumask_empty(mm_cpumask(next)) && switch_mm() 129 !cpumask_test_cpu(cpu, mm_cpumask(next))) switch_mm() 132 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { switch_mm() 133 check_and_switch_context(next, tsk); switch_mm()
|
/linux-4.4.14/arch/frv/include/asm/ |
H A D | switch_to.h | 18 * switch_to(prev, next) should switch from task `prev' to `next' 19 * `prev' will never be the same as `next'. 27 #define switch_to(prev, next, last) \ 31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
H A D | mmu_context.h | 35 #define switch_mm(prev, next, tsk) \ 37 if (prev != next) \ 38 change_mm_context(&prev->context, &next->context, next->pgd); \ 41 #define activate_mm(prev, next) \ 43 change_mm_context(&prev->context, &next->context, next->pgd); \
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | switch_to.h | 26 * @next: The task to begin executing. 27 * @next_ti: task_thread_info(next). 30 * the context of next. Returns prev. 33 struct task_struct *next, struct thread_info *next_ti); 62 next->thread.emulated_fp = 0; \ 84 #define switch_to(prev, next, last) \ 90 __restore_dsp(next); \ 99 if (KSTK_STATUS(next) & ST0_CU2 && \ 101 cop2_restore(next); \ 107 write_c0_userlocal(task_thread_info(next)->tp_value); \ 108 __restore_watch(next); \ 109 (last) = resume(prev, next, task_thread_info(next)); \
|
H A D | mmu_context.h | 139 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 148 if ((cpu_context(cpu, next) ^ asid_cache(cpu)) & ASID_VERSION_MASK) switch_mm() 149 get_new_mmu_context(next, cpu); switch_mm() 150 write_c0_entryhi(cpu_asid(cpu, next)); switch_mm() 151 TLBMISS_HANDLER_SETUP_PGD(next->pgd); switch_mm() 158 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm() 179 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 188 get_new_mmu_context(next, cpu); activate_mm() 190 write_c0_entryhi(cpu_asid(cpu, next)); activate_mm() 191 TLBMISS_HANDLER_SETUP_PGD(next->pgd); activate_mm() 195 cpumask_set_cpu(cpu, mm_cpumask(next)); activate_mm() 217 /* will get a new context next time */ drop_mmu_context()
|
/linux-4.4.14/arch/blackfin/mm/ |
H A D | sram-alloc.c | 31 struct sram_piece *next; member in struct:sram_piece 78 per_cpu(free_l1_ssram_head, cpu).next = l1sram_init() 80 if (!per_cpu(free_l1_ssram_head, cpu).next) { l1sram_init() 85 per_cpu(free_l1_ssram_head, cpu).next->paddr = (void *)get_l1_scratch_start_cpu(cpu) + reserve; l1sram_init() 86 per_cpu(free_l1_ssram_head, cpu).next->size = L1_SCRATCH_LENGTH - reserve; l1sram_init() 87 per_cpu(free_l1_ssram_head, cpu).next->pid = 0; l1sram_init() 88 per_cpu(free_l1_ssram_head, cpu).next->next = NULL; l1sram_init() 90 per_cpu(used_l1_ssram_head, cpu).next = NULL; l1sram_init() 106 per_cpu(free_l1_data_A_sram_head, cpu).next = l1_data_sram_init() 108 if (!per_cpu(free_l1_data_A_sram_head, cpu).next) { l1_data_sram_init() 113 per_cpu(free_l1_data_A_sram_head, cpu).next->paddr = l1_data_sram_init() 115 per_cpu(free_l1_data_A_sram_head, cpu).next->size = l1_data_sram_init() 117 per_cpu(free_l1_data_A_sram_head, cpu).next->pid = 0; l1_data_sram_init() 118 per_cpu(free_l1_data_A_sram_head, cpu).next->next = NULL; l1_data_sram_init() 120 per_cpu(used_l1_data_A_sram_head, cpu).next = NULL; l1_data_sram_init() 124 per_cpu(free_l1_data_A_sram_head, cpu).next->size >> 10); l1_data_sram_init() 129 per_cpu(free_l1_data_B_sram_head, cpu).next = l1_data_sram_init() 131 if (!per_cpu(free_l1_data_B_sram_head, cpu).next) { l1_data_sram_init() 136 per_cpu(free_l1_data_B_sram_head, cpu).next->paddr = l1_data_sram_init() 138 per_cpu(free_l1_data_B_sram_head, cpu).next->size = l1_data_sram_init() 140 per_cpu(free_l1_data_B_sram_head, cpu).next->pid = 0; l1_data_sram_init() 141 per_cpu(free_l1_data_B_sram_head, cpu).next->next = NULL; l1_data_sram_init() 143 per_cpu(used_l1_data_B_sram_head, cpu).next = NULL; l1_data_sram_init() 147 per_cpu(free_l1_data_B_sram_head, cpu).next->size >> 10); l1_data_sram_init() 163 per_cpu(free_l1_inst_sram_head, cpu).next = l1_inst_sram_init() 165 if (!per_cpu(free_l1_inst_sram_head, cpu).next) { l1_inst_sram_init() 170 per_cpu(free_l1_inst_sram_head, cpu).next->paddr = l1_inst_sram_init() 172 per_cpu(free_l1_inst_sram_head, cpu).next->size = l1_inst_sram_init() 174 per_cpu(free_l1_inst_sram_head, cpu).next->pid = 0; l1_inst_sram_init() 175 per_cpu(free_l1_inst_sram_head, cpu).next->next = NULL; l1_inst_sram_init() 177 per_cpu(used_l1_inst_sram_head, cpu).next = NULL; l1_inst_sram_init() 181 per_cpu(free_l1_inst_sram_head, cpu).next->size >> 10); l1_inst_sram_init() 228 free_l2_sram_head.next = l2_sram_init() 230 if (!free_l2_sram_head.next) { l2_sram_init() 235 free_l2_sram_head.next->paddr = l2_sram_init() 237 free_l2_sram_head.next->size = l2_sram_init() 239 free_l2_sram_head.next->pid = 0; l2_sram_init() 240 free_l2_sram_head.next->next = NULL; l2_sram_init() 242 used_l2_sram_head.next = NULL; l2_sram_init() 246 free_l2_sram_head.next->size >> 10); l2_sram_init() 280 pslot = pfree_head->next; _sram_alloc() 286 pslot = pslot->next; _sram_alloc() 293 plast->next = pslot->next; _sram_alloc() 310 pslot = pused_head->next; _sram_alloc() 316 pslot = pslot->next; _sram_alloc() 319 pavail->next = pslot; _sram_alloc() 320 plast->next = pavail; _sram_alloc() 335 pmax = pslot = pfree_head->next; _sram_alloc_max() 341 pslot = pslot->next; _sram_alloc_max() 363 pslot = pused_head->next; _sram_free() 369 pslot = pslot->next; _sram_free() 375 plast->next = pslot->next; _sram_free() 380 pslot = pfree_head->next; _sram_free() 385 pslot = pslot->next; _sram_free() 392 pavail->next = plast->next; _sram_free() 393 plast->next = pavail; _sram_free() 399 plast->next = pslot->next; _sram_free() 746 for (tmp = &mm->context.sram_list; *tmp; tmp = &(*tmp)->next) sram_free_with_lsl() 750 *tmp = lsl->next; sram_free_with_lsl() 791 lsl->next = mm->context.sram_list; sram_alloc_with_lsl() 816 pslot = pused_head->next; _sram_proc_show() 823 pslot = pslot->next; _sram_proc_show() 826 pslot = pfree_head->next; _sram_proc_show() 833 pslot = pslot->next; _sram_proc_show()
|
/linux-4.4.14/net/netlabel/ |
H A D | netlabel_addrlist.h | 78 i = i->next; __af4list_valid() 90 i = rcu_dereference(i->next); __af4list_valid_rcu() 97 for (iter = __af4list_valid((head)->next, head); \ 99 iter = __af4list_valid(iter->list.next, head)) 102 for (iter = __af4list_valid_rcu((head)->next, head); \ 104 iter = __af4list_valid_rcu(iter->list.next, head)) 107 for (iter = __af4list_valid((head)->next, head), \ 108 tmp = __af4list_valid(iter->list.next, head); \ 110 iter = tmp, tmp = __af4list_valid(iter->list.next, head)) 145 i = i->next; __af6list_valid() 157 i = rcu_dereference(i->next); __af6list_valid_rcu() 164 for (iter = __af6list_valid((head)->next, head); \ 166 iter = __af6list_valid(iter->list.next, head)) 169 for (iter = __af6list_valid_rcu((head)->next, head); \ 171 iter = __af6list_valid_rcu(iter->list.next, head)) 174 for (iter = __af6list_valid((head)->next, head), \ 175 tmp = __af6list_valid(iter->list.next, head); \ 177 iter = tmp, tmp = __af6list_valid(iter->list.next, head))
|
/linux-4.4.14/arch/sparc/include/asm/ |
H A D | switch_to_64.h | 6 #define prepare_arch_switch(next) \ 14 * for l0/l1. It will use one for 'next' and the other to hold 15 * the output value of 'last'. 'next' is not referenced again 20 #define switch_to(prev, next, last) \ 25 : : "r" (task_thread_info(next)->current_ds));\ 27 task_thread_info(next); \ 58 : "0" (task_thread_info(next)), \
|
H A D | switch_to_32.h | 27 #define SWITCH_DO_LAZY_FPU(next) /* */ 37 #define prepare_arch_switch(next) do { \ 57 #define switch_to(prev, next, last) do { \ 59 SWITCH_DO_LAZY_FPU(next); \ 60 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next->active_mm)); \ 92 "r" (task_thread_info(next)), \
|
H A D | cpudata_32.h | 24 int next; member in struct:__anon2689
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | switch_to.h | 13 * switch_to(prev, next) should switch from task `prev' to `next' 14 * `prev' will never be the same as `next'. 16 * `next' and `prev' should be struct task_struct, but it isn't always defined 28 #define switch_to(prev, next, last) do { \ 45 "r" (&(prev->thread.sp)), "r" (&(next->thread.sp)), \ 46 "r" (&(prev->thread.lr)), "r" (&(next->thread.lr)) \
|
H A D | mmu_context.h | 122 struct mm_struct *next, struct task_struct *tsk) switch_mm() 128 if (prev != next) { switch_mm() 130 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm() 132 /* Set MPTB = next->pgd */ switch_mm() 133 *(volatile unsigned long *)MPTB = (unsigned long)next->pgd; switch_mm() 134 activate_context(next); switch_mm() 138 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) switch_mm() 139 activate_context(next); switch_mm() 145 #define activate_mm(prev, next) \ 146 switch_mm((prev), (next), NULL) 155 #define switch_mm(prev,next,tsk) do { } while (0) 157 #define activate_mm(prev,next) do { } while (0) 121 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
H A D | mkregtable.c | 36 * sometimes we already know the next/prev entries and we can 42 struct list_head *next, *prev; member in struct:list_head 52 list->next = list; INIT_LIST_HEAD() 60 * the prev/next entries already! 64 struct list_head *prev, struct list_head *next) __list_add() 66 next->prev = new; __list_add() 67 new->next = next; __list_add() 69 prev->next = new; __list_add() 73 struct list_head *prev, struct list_head *next); 86 __list_add(new, head, head->next); list_add() 103 * Delete a list entry by making the prev/next entries 107 * the prev/next entries already! 109 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 111 next->prev = prev; __list_del() 112 prev->next = next; __list_del() 124 __list_del(entry->prev, entry->next); list_del() 125 entry->next = (void *)0xDEADBEEF; list_del() 141 new->next = old->next; list_replace() 142 new->next->prev = new; list_replace() 144 new->prev->next = new; list_replace() 160 __list_del(entry->prev, entry->next); list_del_init() 171 __list_del(list->prev, list->next); list_move() 183 __list_del(list->prev, list->next); list_move_tail() 195 return list->next == head; list_is_last() 204 return head->next == head; list_empty() 213 * in the process of modifying either member (next or prev) 222 struct list_head *next = head->next; list_empty_careful() local 223 return (next == head) && (next == head->prev); list_empty_careful() 232 return !list_empty(head) && (head->next == head->prev); list_is_singular() 239 struct list_head *new_first = entry->next; __list_cut_position() 240 list->next = head->next; __list_cut_position() 241 list->next->prev = list; __list_cut_position() 243 entry->next = list; __list_cut_position() 244 head->next = new_first; __list_cut_position() 268 if (list_is_singular(head) && (head->next != entry && head != entry)) list_cut_position() 277 struct list_head *prev, struct list_head *next) __list_splice() 279 struct list_head *first = list->next; __list_splice() 283 prev->next = first; __list_splice() 285 last->next = next; __list_splice() 286 next->prev = last; __list_splice() 298 __list_splice(list, head, head->next); list_splice() 324 __list_splice(list, head, head->next); list_splice_init() 364 list_entry((ptr)->next, type, member) 372 for (pos = (head)->next; prefetch(pos->next), pos != (head); \ 373 pos = pos->next) 391 for (pos = (head)->next, n = pos->next; pos != (head); \ 392 pos = n, n = pos->next) 412 for (pos = list_entry((head)->next, typeof(*pos), member); \ 414 pos = list_entry(pos->member.next, typeof(*pos), member)) 448 for (pos = list_entry(pos->member.next, typeof(*pos), member); \ 449 prefetch(pos->member.next), &pos->member != (head); \ 450 pos = list_entry(pos->member.next, typeof(*pos), member)) 475 for (; prefetch(pos->member.next), &pos->member != (head); \ 476 pos = list_entry(pos->member.next, typeof(*pos), member)) 486 for (pos = list_entry((head)->next, typeof(*pos), member), \ 487 n = list_entry(pos->member.next, typeof(*pos), member); \ 489 pos = n, n = list_entry(n->member.next, typeof(*n), member)) 502 for (pos = list_entry(pos->member.next, typeof(*pos), member), \ 503 n = list_entry(pos->member.next, typeof(*pos), member); \ 505 pos = n, n = list_entry(n->member.next, typeof(*n), member)) 518 for (n = list_entry(pos->member.next, typeof(*pos), member); \ 520 pos = n, n = list_entry(n->member.next, typeof(*n), member)) 63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 276 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
|
H A D | radeon_mem.c | 16 * The above copyright notice and this permission notice (including the next 54 newblock->next = p->next; split_block() 56 p->next->prev = newblock; split_block() 57 p->next = newblock; split_block() 71 newblock->next = p->next; split_block() 73 p->next->prev = newblock; split_block() 74 p->next = newblock; split_block() 117 if (p->next->file_priv == NULL) { free_block() 118 struct mem_block *q = p->next; free_block() 120 p->next = q->next; free_block() 121 p->next->prev = p; free_block() 128 q->next = p->next; free_block() 129 q->next->prev = q; free_block() 152 blocks->next = blocks->prev = *heap; init_heap() 155 (*heap)->next = (*heap)->prev = blocks; init_heap() 165 if (!heap || !heap->next) radeon_mem_release() 177 while (p->file_priv == NULL && p->next->file_priv == NULL) { list_for_each() 178 struct mem_block *q = p->next; list_for_each() 180 p->next = q->next; list_for_each() 181 p->next->prev = p; list_for_each() 196 for (p = (*heap)->next; p != *heap;) { radeon_mem_takedown() 198 p = p->next; radeon_mem_takedown()
|
/linux-4.4.14/security/selinux/ss/ |
H A D | hashtab.c | 54 cur = cur->next; hashtab_insert() 66 newnode->next = prev->next; hashtab_insert() 67 prev->next = newnode; hashtab_insert() 69 newnode->next = h->htable[hvalue]; hashtab_insert() 88 cur = cur->next; hashtab_search() 108 cur = cur->next; hashtab_destroy() 137 cur = cur->next; hashtab_map() 158 cur = cur->next; hashtab_stat()
|
H A D | ebitmap.c | 39 n1 = n1->next; ebitmap_cmp() 40 n2 = n2->next; ebitmap_cmp() 64 new->next = NULL; ebitmap_cpy() 66 prev->next = new; ebitmap_cpy() 70 n = n->next; ebitmap_cpy() 120 e_iter = e_iter->next; ebitmap_netlbl_export() 172 e_prev->next = e_iter; ebitmap_netlbl_import() 180 /* next */ ebitmap_netlbl_import() 211 n1 = n1->next; ebitmap_contains() 229 n1 = n1->next; ebitmap_contains() 230 n2 = n2->next; ebitmap_contains() 250 n = n->next; ebitmap_get_bit() 276 if (!n->next) { ebitmap_set_bit() 288 prev->next = n->next; ebitmap_set_bit() 290 e->node = n->next; ebitmap_set_bit() 296 n = n->next; ebitmap_set_bit() 314 new->next = prev->next; ebitmap_set_bit() 315 prev->next = new; ebitmap_set_bit() 317 new->next = e->node; ebitmap_set_bit() 334 n = n->next; ebitmap_destroy() 410 n->next = tmp; ebitmap_read() 496 /* set up for the next node */ ebitmap_for_each_positive_bit()
|
H A D | sidtab.c | 48 cur = cur->next; sidtab_insert() 69 newnode->next = prev->next; sidtab_insert() 71 prev->next = newnode; sidtab_insert() 73 newnode->next = s->htable[hvalue]; sidtab_insert() 96 cur = cur->next; sidtab_search_core() 107 cur = cur->next; sidtab_search_core() 143 cur = cur->next; sidtab_map() 174 cur = cur->next; sidtab_search_context() 254 cur = cur->next; sidtab_hash_eval() 279 cur = cur->next; sidtab_destroy()
|
H A D | constraint.h | 53 struct constraint_expr *next; /* next expression */ member in struct:constraint_expr 59 struct constraint_node *next; /* next constraint */ member in struct:constraint_node
|
H A D | sidtab.h | 15 struct sidtab_node *next; member in struct:sidtab_node 27 unsigned int next_sid; /* next SID to allocate */
|
H A D | conditional.c | 33 for (cur = expr; cur; cur = cur->next) { cond_evaluate_expr() 101 for (cur = node->true_list; cur; cur = cur->next) { evaluate_cond_node() 108 for (cur = node->false_list; cur; cur = cur->next) { evaluate_cond_node() 135 struct cond_av_list *cur, *next; cond_av_list_destroy() local 136 for (cur = list; cur; cur = next) { cond_av_list_destroy() 137 next = cur->next; cond_av_list_destroy() 148 next_expr = cur_expr->next; cond_node_destroy() 158 struct cond_node *next, *cur; cond_list_destroy() local 163 for (cur = list; cur; cur = next) { cond_list_destroy() 164 next = cur->next; cond_list_destroy() 306 for (cur = other; cur; cur = cur->next) { cond_insertf() 342 data->tail->next = list; cond_insertf() 437 last->next = expr; cond_read_node() 483 last->next = node; cond_read_list() 534 for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) cond_write_av_list() 545 for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) { cond_write_av_list() 567 for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) cond_write_node() 575 for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) { cond_write_node() 601 for (cur = list; cur != NULL; cur = cur->next) cond_write_list() 608 for (cur = list; cur != NULL; cur = cur->next) { cond_write_list()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | gddr3.c | 75 switch (ram->next->bios.timing_ver) { nvkm_gddr3_calc() 77 CWL = ram->next->bios.timing_10_CWL; nvkm_gddr3_calc() 78 CL = ram->next->bios.timing_10_CL; nvkm_gddr3_calc() 79 WR = ram->next->bios.timing_10_WR; nvkm_gddr3_calc() 80 DLL = !ram->next->bios.ramcfg_DLLoff; nvkm_gddr3_calc() 81 ODT = ram->next->bios.timing_10_ODT; nvkm_gddr3_calc() 82 RON = ram->next->bios.ramcfg_RON; nvkm_gddr3_calc() 85 CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; nvkm_gddr3_calc() 86 CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; nvkm_gddr3_calc() 87 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_gddr3_calc() 96 if (ram->next->bios.timing_ver == 0x20 || nvkm_gddr3_calc() 97 ram->next->bios.ramcfg_timing == 0xff) { nvkm_gddr3_calc()
|
H A D | gddr5.c | 41 xd = !ram->next->bios.ramcfg_DLLoff; nvkm_gddr5_calc() 43 switch (ram->next->bios.ramcfg_ver) { nvkm_gddr5_calc() 45 pd = ram->next->bios.ramcfg_11_01_80; nvkm_gddr5_calc() 46 lf = ram->next->bios.ramcfg_11_01_40; nvkm_gddr5_calc() 47 vh = ram->next->bios.ramcfg_11_02_10; nvkm_gddr5_calc() 48 vr = ram->next->bios.ramcfg_11_02_04; nvkm_gddr5_calc() 49 vo = ram->next->bios.ramcfg_11_06; nvkm_gddr5_calc() 50 l3 = !ram->next->bios.ramcfg_11_07_02; nvkm_gddr5_calc() 56 switch (ram->next->bios.timing_ver) { nvkm_gddr5_calc() 58 WL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; nvkm_gddr5_calc() 59 CL = (ram->next->bios.timing[1] & 0x0000001f); nvkm_gddr5_calc() 60 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_gddr5_calc() 61 at[0] = ram->next->bios.timing_20_2e_c0; nvkm_gddr5_calc() 62 at[1] = ram->next->bios.timing_20_2e_30; nvkm_gddr5_calc() 63 dt = ram->next->bios.timing_20_2e_03; nvkm_gddr5_calc() 64 ds = ram->next->bios.timing_20_2f_03; nvkm_gddr5_calc()
|
H A D | sddr3.c | 73 DLL = !ram->next->bios.ramcfg_DLLoff; nvkm_sddr3_calc() 75 switch (ram->next->bios.timing_ver) { nvkm_sddr3_calc() 77 if (ram->next->bios.timing_hdr < 0x17) { nvkm_sddr3_calc() 81 CWL = ram->next->bios.timing_10_CWL; nvkm_sddr3_calc() 82 CL = ram->next->bios.timing_10_CL; nvkm_sddr3_calc() 83 WR = ram->next->bios.timing_10_WR; nvkm_sddr3_calc() 84 ODT = ram->next->bios.timing_10_ODT; nvkm_sddr3_calc() 87 CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; nvkm_sddr3_calc() 88 CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; nvkm_sddr3_calc() 89 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_sddr3_calc()
|
H A D | ramgk104.c | 210 struct nvkm_ram_data *next = ram->base.next; r1373f4_fini() local 211 u8 v0 = next->bios.ramcfg_11_03_c0; r1373f4_fini() 212 u8 v1 = next->bios.ramcfg_11_03_30; r1373f4_fini() 243 u32 next = (prev & ~mask) | data; gk104_ram_nuts() local 244 nvkm_memx_wr32(fuc->memx, addr, next); gk104_ram_nuts() 255 struct nvkm_ram_data *next = ram->base.next; gk104_ram_calc_gddr5() local 256 int vc = !next->bios.ramcfg_11_02_08; gk104_ram_calc_gddr5() 257 int mv = !next->bios.ramcfg_11_02_04; gk104_ram_calc_gddr5() 309 switch (next->bios.ramcfg_11_03_c0) { gk104_ram_calc_gddr5() 316 switch (next->bios.ramcfg_11_03_30) { gk104_ram_calc_gddr5() 324 if (next->bios.ramcfg_11_02_80) gk104_ram_calc_gddr5() 326 if (next->bios.ramcfg_11_02_40) gk104_ram_calc_gddr5() 328 if (next->bios.ramcfg_11_07_10) gk104_ram_calc_gddr5() 330 if (next->bios.ramcfg_11_07_08) gk104_ram_calc_gddr5() 365 if (next->bios.ramcfg_11_02_40 || gk104_ram_calc_gddr5() 366 next->bios.ramcfg_11_07_10) { gk104_ram_calc_gddr5() 389 if (next->bios.ramcfg_11_07_40) gk104_ram_calc_gddr5() 393 ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c); gk104_ram_calc_gddr5() 394 ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_gddr5() 395 ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_gddr5() 397 if (!next->bios.ramcfg_11_07_08 && !next->bios.ramcfg_11_07_04) { gk104_ram_calc_gddr5() 398 ram_wr32(fuc, 0x10f698, 0x01010101 * next->bios.ramcfg_11_04); gk104_ram_calc_gddr5() 399 ram_wr32(fuc, 0x10f69c, 0x01010101 * next->bios.ramcfg_11_04); gk104_ram_calc_gddr5() 401 if (!next->bios.ramcfg_11_07_08) { gk104_ram_calc_gddr5() 407 u32 data = 0x01000100 * next->bios.ramcfg_11_04; gk104_ram_calc_gddr5() 412 if (ram->mode == 2 && next->bios.ramcfg_11_08_10) gk104_ram_calc_gddr5() 420 if (!next->bios.ramcfg_11_02_80) gk104_ram_calc_gddr5() 422 if (!next->bios.ramcfg_11_02_40) gk104_ram_calc_gddr5() 424 if (!next->bios.ramcfg_11_07_10) gk104_ram_calc_gddr5() 426 if (!next->bios.ramcfg_11_07_08) gk104_ram_calc_gddr5() 432 if (next->bios.ramcfg_11_01_08) gk104_ram_calc_gddr5() 443 if (next->bios.ramcfg_11_08_01) gk104_ram_calc_gddr5() 450 if (next->bios.ramcfg_11_08_08) gk104_ram_calc_gddr5() 452 if (next->bios.ramcfg_11_08_04) gk104_ram_calc_gddr5() 454 if (next->bios.ramcfg_11_08_02) gk104_ram_calc_gddr5() 459 ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]); gk104_ram_calc_gddr5() 460 ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]); gk104_ram_calc_gddr5() 461 ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]); gk104_ram_calc_gddr5() 462 ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]); gk104_ram_calc_gddr5() 463 ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]); gk104_ram_calc_gddr5() 464 ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]); gk104_ram_calc_gddr5() 465 ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]); gk104_ram_calc_gddr5() 466 ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]); gk104_ram_calc_gddr5() 467 ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]); gk104_ram_calc_gddr5() 468 ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]); gk104_ram_calc_gddr5() 469 ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]); gk104_ram_calc_gddr5() 473 if (next->bios.ramcfg_11_08_20) gk104_ram_calc_gddr5() 481 data |= next->bios.ramcfg_11_02_03 << 8; gk104_ram_calc_gddr5() 485 if (next->bios.ramcfg_11_01_10) gk104_ram_calc_gddr5() 493 data |= next->bios.timing_20_30_07 << 28; gk104_ram_calc_gddr5() 497 if (next->bios.ramcfg_11_01_01) gk104_ram_calc_gddr5() 505 data |= next->bios.timing_20_30_07 << 28; gk104_ram_calc_gddr5() 509 if (next->bios.ramcfg_11_01_02) gk104_ram_calc_gddr5() 517 if (!next->bios.ramcfg_11_01_04) gk104_ram_calc_gddr5() 519 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_gddr5() 524 if (next->bios.ramcfg_11_03_f0) { gk104_ram_calc_gddr5() 525 if (next->bios.rammap_11_08_0c) { gk104_ram_calc_gddr5() 526 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_gddr5() 539 ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f); gk104_ram_calc_gddr5() 543 data |= next->bios.ramcfg_11_02_03; gk104_ram_calc_gddr5() 547 if (next->bios.ramcfg_11_01_10) gk104_ram_calc_gddr5() 558 data = next->bios.timing_20_30_07 << 8; gk104_ram_calc_gddr5() 559 if (next->bios.ramcfg_11_01_01) gk104_ram_calc_gddr5() 563 ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4); gk104_ram_calc_gddr5() 564 data = (next->bios.timing[10] & 0x7f000000) >> 24; gk104_ram_calc_gddr5() 565 if (data < next->bios.timing_20_2c_1fc0) gk104_ram_calc_gddr5() 566 data = next->bios.timing_20_2c_1fc0; gk104_ram_calc_gddr5() 568 ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16); gk104_ram_calc_gddr5() 570 ram_mask(fuc, 0x10fec4, 0x041e0f07, next->bios.timing_20_31_0800 << 26 | gk104_ram_calc_gddr5() 571 next->bios.timing_20_31_0780 << 17 | gk104_ram_calc_gddr5() 572 next->bios.timing_20_31_0078 << 8 | gk104_ram_calc_gddr5() 573 next->bios.timing_20_31_0007); gk104_ram_calc_gddr5() 574 ram_mask(fuc, 0x10fec8, 0x00000027, next->bios.timing_20_31_8000 << 5 | gk104_ram_calc_gddr5() 575 next->bios.timing_20_31_7000); gk104_ram_calc_gddr5() 583 if (next->bios.ramcfg_11_08_10 && (ram->mode == 2) /*XXX*/) { gk104_ram_calc_gddr5() 616 if (!next->bios.ramcfg_11_07_08) { gk104_ram_calc_gddr5() 617 if (!next->bios.ramcfg_11_07_04) gk104_ram_calc_gddr5() 631 if (!next->bios.ramcfg_11_07_08) { gk104_ram_calc_gddr5() 633 if ( next->bios.ramcfg_11_07_04) gk104_ram_calc_gddr5() 635 if (!next->bios.rammap_11_08_10) gk104_ram_calc_gddr5() 657 if (next->bios.ramcfg_11_07_02) gk104_ram_calc_gddr5() 663 if (next->bios.rammap_11_08_01) gk104_ram_calc_gddr5() 702 struct nvkm_ram_data *next = ram->base.next; gk104_ram_calc_sddr3() local 703 int vc = !next->bios.ramcfg_11_02_08; gk104_ram_calc_sddr3() 704 int mv = !next->bios.ramcfg_11_02_04; gk104_ram_calc_sddr3() 720 if (next->bios.ramcfg_11_03_f0) gk104_ram_calc_sddr3() 725 if (next->bios.ramcfg_DLLoff) gk104_ram_calc_sddr3() 749 switch (next->bios.ramcfg_11_03_c0) { gk104_ram_calc_sddr3() 756 switch (next->bios.ramcfg_11_03_30) { gk104_ram_calc_sddr3() 764 if (next->bios.ramcfg_11_02_80) gk104_ram_calc_sddr3() 766 if (next->bios.ramcfg_11_02_40) gk104_ram_calc_sddr3() 768 if (next->bios.ramcfg_11_07_10) gk104_ram_calc_sddr3() 770 if (next->bios.ramcfg_11_07_08) gk104_ram_calc_sddr3() 780 data |= next->bios.ramcfg_11_03_30 << 16; gk104_ram_calc_sddr3() 812 if (next->bios.ramcfg_11_02_40 || gk104_ram_calc_sddr3() 813 next->bios.ramcfg_11_07_10) { gk104_ram_calc_sddr3() 819 if (next->bios.ramcfg_11_07_40) gk104_ram_calc_sddr3() 823 ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c); gk104_ram_calc_sddr3() 824 ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_sddr3() 825 ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_sddr3() 829 if (!next->bios.ramcfg_11_02_80) gk104_ram_calc_sddr3() 831 if (!next->bios.ramcfg_11_02_40) gk104_ram_calc_sddr3() 833 if (!next->bios.ramcfg_11_07_10) gk104_ram_calc_sddr3() 835 if (!next->bios.ramcfg_11_07_08) gk104_ram_calc_sddr3() 842 if (next->bios.ramcfg_11_08_01) gk104_ram_calc_sddr3() 849 ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]); gk104_ram_calc_sddr3() 850 ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]); gk104_ram_calc_sddr3() 851 ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]); gk104_ram_calc_sddr3() 852 ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]); gk104_ram_calc_sddr3() 853 ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]); gk104_ram_calc_sddr3() 854 ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]); gk104_ram_calc_sddr3() 855 ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]); gk104_ram_calc_sddr3() 856 ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]); gk104_ram_calc_sddr3() 857 ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]); gk104_ram_calc_sddr3() 858 ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]); gk104_ram_calc_sddr3() 859 ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]); gk104_ram_calc_sddr3() 863 if (!next->bios.ramcfg_11_01_04) gk104_ram_calc_sddr3() 865 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_sddr3() 870 if (next->bios.ramcfg_11_03_f0) { gk104_ram_calc_sddr3() 871 if (next->bios.rammap_11_08_0c) { gk104_ram_calc_sddr3() 872 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_sddr3() 886 ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f); gk104_ram_calc_sddr3() 888 ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4); gk104_ram_calc_sddr3() 890 data = (next->bios.timing[10] & 0x7f000000) >> 24; gk104_ram_calc_sddr3() 891 if (data < next->bios.timing_20_2c_1fc0) gk104_ram_calc_sddr3() 892 data = next->bios.timing_20_2c_1fc0; gk104_ram_calc_sddr3() 895 ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16); gk104_ram_calc_sddr3() 905 if (!next->bios.ramcfg_DLLoff) { gk104_ram_calc_sddr3() 915 if (!next->bios.ramcfg_DLLoff) { gk104_ram_calc_sddr3() 941 if (next->bios.rammap_11_08_01) gk104_ram_calc_sddr3() 1035 gk104_ram_calc_xits(struct gk104_ram *ram, struct nvkm_ram_data *next) gk104_ram_calc_xits() argument 1046 ram->mode = (next->freq > fuc->refpll.vco1.max_freq) ? 2 : 1; gk104_ram_calc_xits() 1057 refclk = next->freq; gk104_ram_calc_xits() 1059 ret = gk104_pll_calc_hiclk(next->freq, subdev->device->crystal, gk104_ram_calc_xits() 1068 " (refclock: %i kHz)\n", next->freq, ret); gk104_ram_calc_xits() 1084 ram->base.freq = next->freq; gk104_ram_calc_xits() 1090 ret = gk104_ram_calc_sddr3(ram, next->freq); gk104_ram_calc_xits() 1095 ret = gk104_ram_calc_gddr5(ram, next->freq); gk104_ram_calc_xits() 1114 if (ram->base.next == NULL) { gk104_ram_calc() 1137 ram->base.next = &ram->base.target; gk104_ram_calc() 1139 ram->base.next = &ram->base.xition; gk104_ram_calc() 1141 BUG_ON(ram->base.next != &ram->base.xition); gk104_ram_calc() 1142 ram->base.next = &ram->base.target; gk104_ram_calc() 1145 return gk104_ram_calc_xits(ram, ram->base.next); gk104_ram_calc() 1230 struct nvkm_ram_data *next = ram->base.next; gk104_ram_prog() local 1234 return (ram->base.next == &ram->base.xition); gk104_ram_prog() 1239 gk104_ram_prog_0(ram, next->freq); gk104_ram_prog() 1241 return (ram->base.next == &ram->base.xition); gk104_ram_prog() 1248 ram->base.next = NULL; gk104_ram_tidy()
|
H A D | sddr2.c | 64 switch (ram->next->bios.timing_ver) { nvkm_sddr2_calc() 66 CL = ram->next->bios.timing_10_CL; nvkm_sddr2_calc() 67 WR = ram->next->bios.timing_10_WR; nvkm_sddr2_calc() 68 DLL = !ram->next->bios.ramcfg_DLLoff; nvkm_sddr2_calc() 69 ODT = ram->next->bios.timing_10_ODT & 3; nvkm_sddr2_calc() 72 CL = (ram->next->bios.timing[1] & 0x0000001f); nvkm_sddr2_calc() 73 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_sddr2_calc() 79 if (ram->next->bios.timing_ver == 0x20 || nvkm_sddr2_calc() 80 ram->next->bios.ramcfg_timing == 0xff) { nvkm_sddr2_calc()
|
H A D | ramnv50.c | 97 unkt3b = 0x19 + ram->base.next->bios.rammap_00_16_40; nv50_ram_timing_calc() 99 ram->base.next->bios.rammap_00_16_40) << 16 | nv50_ram_timing_calc() 227 struct nvkm_ram_data *next; nv50_ram_calc() local 235 next = &ram->base.target; nv50_ram_calc() 236 next->freq = freq; nv50_ram_calc() 237 ram->base.next = next; nv50_ram_calc() 251 nvbios_rammapEp_from_perf(bios, data, hdr, &next->bios); nv50_ram_calc() 261 &next->bios); nv50_ram_calc() 268 if (next->bios.ramcfg_timing != 0xff) { nv50_ram_calc() 269 data = nvbios_timingEp(bios, next->bios.ramcfg_timing, nv50_ram_calc() 270 &ver, &hdr, &cnt, &len, &next->bios); nv50_ram_calc() 305 if (subdev->device->chipset <= 0x96 && !next->bios.ramcfg_00_03_02) nv50_ram_calc() 319 if (next->bios.timing_10_ODT) nv50_ram_calc() 355 next->bios.rammap_00_16_40 << 14); nv50_ram_calc() 363 nv50_ram_gpio(hwsq, 0x18, !next->bios.ramcfg_FBVDDQ); nv50_ram_calc() 400 if (!next->bios.ramcfg_00_03_02) nv50_ram_calc() 402 ram_mask(hwsq, 0x100200, 0x00001000, !next->bios.ramcfg_00_04_02 << 12); nv50_ram_calc() 413 if (!next->bios.ramcfg_00_03_08) nv50_ram_calc() 415 if (!next->bios.ramcfg_FBVDDQ) nv50_ram_calc() 417 if ( next->bios.ramcfg_00_04_04) nv50_ram_calc() 422 if (!next->bios.ramcfg_00_03_08) nv50_ram_calc() 426 if ( next->bios.ramcfg_00_03_01) nv50_ram_calc() 428 if ( next->bios.ramcfg_00_03_02) nv50_ram_calc() 430 if (!next->bios.ramcfg_00_03_08) nv50_ram_calc() 432 if ( next->bios.ramcfg_00_04_04) nv50_ram_calc() 434 if ( next->bios.ramcfg_00_04_20) nv50_ram_calc() 444 if (next->bios.rammap_00_16_20) { nv50_ram_calc() 445 ram_wr32(hwsq, 0x1005a0, next->bios.ramcfg_00_07 << 16 | nv50_ram_calc() 446 next->bios.ramcfg_00_06 << 8 | nv50_ram_calc() 447 next->bios.ramcfg_00_05); nv50_ram_calc() 448 ram_wr32(hwsq, 0x1005a4, next->bios.ramcfg_00_09 << 8 | nv50_ram_calc() 449 next->bios.ramcfg_00_08); nv50_ram_calc() 456 if (!next->bios.timing_10_ODT) nv50_ram_calc() 460 if (!next->bios.ramcfg_DLLoff) nv50_ram_calc() 468 if (next->bios.rammap_00_17_02) nv50_ram_calc() 470 if (!next->bios.rammap_00_16_40) nv50_ram_calc() 472 if (next->bios.ramcfg_00_03_02) nv50_ram_calc() 474 if (subdev->device->chipset <= 0x96 && next->bios.ramcfg_00_03_02) nv50_ram_calc()
|
/linux-4.4.14/arch/c6x/include/asm/ |
H A D | switch_to.h | 21 struct thread_struct *next, 24 #define switch_to(prev, next, last) \ 28 &(next)->thread, (prev)); \
|
/linux-4.4.14/mm/ |
H A D | pagewalk.c | 31 unsigned long next; walk_pmd_range() local 37 next = pmd_addr_end(addr, end); walk_pmd_range() 40 err = walk->pte_hole(addr, next, walk); walk_pmd_range() 50 err = walk->pmd_entry(pmd, addr, next, walk); walk_pmd_range() 64 err = walk_pte_range(pmd, addr, next, walk); walk_pmd_range() 67 } while (pmd++, addr = next, addr != end); walk_pmd_range() 76 unsigned long next; walk_pud_range() local 81 next = pud_addr_end(addr, end); walk_pud_range() 84 err = walk->pte_hole(addr, next, walk); walk_pud_range() 90 err = walk_pmd_range(pud, addr, next, walk); walk_pud_range() 93 } while (pud++, addr = next, addr != end); walk_pud_range() 102 unsigned long next; walk_pgd_range() local 107 next = pgd_addr_end(addr, end); walk_pgd_range() 110 err = walk->pte_hole(addr, next, walk); walk_pgd_range() 116 err = walk_pud_range(pgd, addr, next, walk); walk_pgd_range() 119 } while (pgd++, addr = next, addr != end); walk_pgd_range() 137 unsigned long next; walk_hugetlb_range() local 143 next = hugetlb_entry_end(h, addr, end); walk_hugetlb_range() 146 err = walk->hugetlb_entry(pte, hmask, addr, next, walk); walk_hugetlb_range() 149 } while (addr = next, addr != end); walk_hugetlb_range() 243 unsigned long next; walk_page_range() local 258 next = end; walk_page_range() 261 next = min(end, vma->vm_start); walk_page_range() 264 next = min(end, vma->vm_end); walk_page_range() 267 err = walk_page_test(start, next, walk); walk_page_range() 281 err = __walk_page_range(start, next, walk); walk_page_range() 284 } while (start = next, start < end); walk_page_range()
|
H A D | slob.c | 22 * sufficient free blocks (using a next-fit-like approach) followed by 78 * or offset of next block if -ve (in SLOB_UNITs). 80 * Free blocks of size 1 unit simply contain the offset of the next block. 82 * memory, and the offset of the next free block in the second SLOB_UNIT. 143 * Encode the given size and next info into a free slob block s. 145 static void set_slob(slob_t *s, slobidx_t size, slob_t *next) set_slob() argument 148 slobidx_t offset = next - base; set_slob() 168 * Return the next free slob block pointer after this one. 173 slobidx_t next; slob_next() local 176 next = -s[0].units; slob_next() 178 next = s[1].units; slob_next() 179 return base+next; slob_next() 230 slob_t *next; slob_page_alloc() local 233 next = slob_next(cur); slob_page_alloc() 234 set_slob(aligned, avail - delta, next); slob_page_alloc() 241 next = slob_next(cur); slob_page_alloc() 244 set_slob(prev, slob_units(prev), next); slob_page_alloc() 246 sp->freelist = next; slob_page_alloc() 252 set_slob(cur + units, avail - units, next); slob_page_alloc() 305 * search time by starting our next search here. (see list_for_each_entry() 308 slob_list->next != prev->next) list_for_each_entry() 309 list_move_tail(slob_list, prev->next); list_for_each_entry() 343 slob_t *prev, *next, *b = (slob_t *)block; slob_free() local 400 next = slob_next(prev); slob_free() 401 while (b > next) { slob_free() 402 prev = next; slob_free() 403 next = slob_next(prev); slob_free() 406 if (!slob_last(prev) && b + units == next) { slob_free() 407 units += slob_units(next); slob_free() 408 set_slob(b, units, slob_next(next)); slob_free() 410 set_slob(b, units, next); slob_free()
|
/linux-4.4.14/include/linux/ |
H A D | timerqueue.h | 15 struct timerqueue_node *next; member in struct:timerqueue_head 37 return head->next; timerqueue_getnext() 48 head->next = NULL; timerqueue_init_head()
|
H A D | user-return-notifier.h | 19 struct task_struct *next) propagate_user_return_notify() 23 set_tsk_thread_flag(next, TIF_USER_RETURN_NOTIFY); propagate_user_return_notify() 39 struct task_struct *next) propagate_user_return_notify() 18 propagate_user_return_notify(struct task_struct *prev, struct task_struct *next) propagate_user_return_notify() argument 38 propagate_user_return_notify(struct task_struct *prev, struct task_struct *next) propagate_user_return_notify() argument
|
H A D | list_bl.h | 38 struct hlist_bl_node *next, **pprev; member in struct:hlist_bl_node 45 h->next = NULL; INIT_HLIST_BL_NODE() 81 n->next = first; hlist_bl_add_head() 83 first->pprev = &n->next; hlist_bl_add_head() 90 struct hlist_bl_node *next = n->next; __hlist_bl_del() local 98 ((unsigned long)next | __hlist_bl_del() 100 if (next) __hlist_bl_del() 101 next->pprev = pprev; __hlist_bl_del() 107 n->next = LIST_POISON1; hlist_bl_del() 146 pos = pos->next) 158 pos && ({ n = pos->next; 1; }) && \
|
H A D | list_nulls.h | 25 struct hlist_nulls_node *next, **pprev; member in struct:hlist_nulls_node 68 n->next = first; hlist_nulls_add_head() 72 first->pprev = &n->next; hlist_nulls_add_head() 77 struct hlist_nulls_node *next = n->next; __hlist_nulls_del() local 80 WRITE_ONCE(*pprev, next); __hlist_nulls_del() 81 if (!is_a_nulls(next)) __hlist_nulls_del() 82 next->pprev = pprev; __hlist_nulls_del() 103 pos = pos->next) 115 pos = pos->next)
|
H A D | list.h | 15 * sometimes we already know the next/prev entries and we can 27 list->next = list; INIT_LIST_HEAD() 35 * the prev/next entries already! 40 struct list_head *next) __list_add() 42 next->prev = new; __list_add() 43 new->next = next; __list_add() 45 prev->next = new; __list_add() 50 struct list_head *next); 63 __list_add(new, head, head->next); list_add() 81 * Delete a list entry by making the prev/next entries 85 * the prev/next entries already! 87 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument 89 next->prev = prev; __list_del() 90 WRITE_ONCE(prev->next, next); __list_del() 102 __list_del(entry->prev, entry->next); __list_del_entry() 107 __list_del(entry->prev, entry->next); list_del() 108 entry->next = LIST_POISON1; list_del() 126 new->next = old->next; list_replace() 127 new->next->prev = new; list_replace() 129 new->prev->next = new; list_replace() 180 return list->next == head; list_is_last() 189 return head->next == head; list_empty() 198 * in the process of modifying either member (next or prev) 207 struct list_head *next = head->next; list_empty_careful() local 208 return (next == head) && (next == head->prev); list_empty_careful() 220 first = head->next; list_rotate_left() 231 return !list_empty(head) && (head->next == head->prev); list_is_singular() 237 struct list_head *new_first = entry->next; __list_cut_position() 238 list->next = head->next; __list_cut_position() 239 list->next->prev = list; __list_cut_position() 241 entry->next = list; __list_cut_position() 242 head->next = new_first; __list_cut_position() 266 (head->next != entry && head != entry)) list_cut_position() 276 struct list_head *next) __list_splice() 278 struct list_head *first = list->next; __list_splice() 282 prev->next = first; __list_splice() 284 last->next = next; __list_splice() 285 next->prev = last; __list_splice() 297 __list_splice(list, head, head->next); list_splice() 323 __list_splice(list, head, head->next); list_splice_init() 363 list_entry((ptr)->next, type, member) 388 * list_next_entry - get the next element in list 393 list_entry((pos)->member.next, typeof(*(pos)), member) 409 for (pos = (head)->next; pos != (head); pos = pos->next) 426 for (pos = (head)->next, n = pos->next; pos != (head); \ 427 pos = n, n = pos->next) 600 h->next = NULL; INIT_HLIST_NODE() 616 struct hlist_node *next = n->next; __hlist_del() local 619 WRITE_ONCE(*pprev, next); __hlist_del() 620 if (next) __hlist_del() 621 next->pprev = pprev; __hlist_del() 627 n->next = LIST_POISON1; hlist_del() 642 n->next = first; hlist_add_head() 644 first->pprev = &n->next; hlist_add_head() 649 /* next must be != NULL */ hlist_add_before() 651 struct hlist_node *next) hlist_add_before() 653 n->pprev = next->pprev; hlist_add_before() 654 n->next = next; hlist_add_before() 655 next->pprev = &n->next; hlist_add_before() 662 n->next = prev->next; hlist_add_behind() 663 prev->next = n; hlist_add_behind() 664 n->pprev = &prev->next; hlist_add_behind() 666 if (n->next) hlist_add_behind() 667 n->next->pprev = &n->next; hlist_add_behind() 673 n->pprev = &n->next; hlist_add_fake() 678 return h->pprev == &h->next; hlist_fake() 697 for (pos = (head)->first; pos ; pos = pos->next) 700 for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \ 717 pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) 725 for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\ 727 pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) 736 pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) 747 pos && ({ n = pos->member.next; 1; }); \ 38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument 650 hlist_add_before(struct hlist_node *n, struct hlist_node *next) hlist_add_before() argument
|
H A D | rculist.h | 32 WRITE_ONCE(list->next, list); INIT_LIST_HEAD_RCU() 37 * return the ->next pointer of a list_head in an rcu safe 40 #define list_next_rcu(list) (*((struct list_head __rcu **)(&(list)->next))) 46 * the prev/next entries already! 50 struct list_head *prev, struct list_head *next) __list_add_rcu() 52 new->next = next; __list_add_rcu() 55 next->prev = new; __list_add_rcu() 59 struct list_head *prev, struct list_head *next); 80 __list_add_rcu(new, head, head->next); list_add_rcu() 174 new->next = old->next; list_replace_rcu() 177 new->next->prev = new; list_replace_rcu() 202 struct list_head *first = list->next; list_splice_init_rcu() 204 struct list_head *at = head->next; list_splice_init_rcu() 234 last->next = at; list_splice_init_rcu() 265 * list_first_entry_rcu rereads the ->next pointer. 267 * Rereading the ->next pointer is not a problem for list_empty() and 288 struct list_head *__next = READ_ONCE(__ptr->next); \ 303 for (pos = list_entry_rcu((head)->next, typeof(*pos), member); \ 305 pos = list_entry_rcu(pos->member.next, typeof(*pos), member)) 317 for (pos = list_entry_rcu(pos->member.next, typeof(*pos), member); \ 319 pos = list_entry_rcu(pos->member.next, typeof(*pos), member)) 356 struct hlist_node *next = old->next; hlist_replace_rcu() local 358 new->next = next; hlist_replace_rcu() 361 if (next) hlist_replace_rcu() 362 new->next->pprev = &new->next; hlist_replace_rcu() 367 * return the first or the next element in an RCU protected hlist 370 #define hlist_next_rcu(node) (*((struct hlist_node __rcu **)(&(node)->next))) 397 n->next = first; hlist_add_head_rcu() 401 first->pprev = &n->next; hlist_add_head_rcu() 407 * @next: the existing element to add the new element before. 423 struct hlist_node *next) hlist_add_before_rcu() 425 n->pprev = next->pprev; hlist_add_before_rcu() 426 n->next = next; hlist_add_before_rcu() 428 next->pprev = &n->next; hlist_add_before_rcu() 452 n->next = prev->next; hlist_add_behind_rcu() 453 n->pprev = &prev->next; hlist_add_behind_rcu() 455 if (n->next) hlist_add_behind_rcu() 456 n->next->pprev = &n->next; hlist_add_behind_rcu() 49 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument 422 hlist_add_before_rcu(struct hlist_node *n, struct hlist_node *next) hlist_add_before_rcu() argument
|
H A D | if_tunnel.h | 14 for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
|
H A D | ihex.h | 23 /* Find the next record, taking into account the 4-byte alignment */ 27 int next = ((be16_to_cpu(rec->len) + 5) & ~3) - 2; ihex_next_binrec() local 28 rec = (void *)&rec->data[next]; ihex_next_binrec() 46 /* Point to next record... */ ihex_validate_fw()
|
H A D | initrd.h | 16 /* free_initrd_mem always gets called with the next two as arguments.. */
|
/linux-4.4.14/lib/lzo/ |
H A D | lzo1x_decompress_safe.c | 43 size_t t, next; lzo1x_decompress_safe() local 57 next = t; lzo1x_decompress_safe() 110 next = t & 3; lzo1x_decompress_safe() 121 next = t & 3; lzo1x_decompress_safe() 128 next = t & 3; lzo1x_decompress_safe() 152 next = get_unaligned_le16(ip); lzo1x_decompress_safe() 154 m_pos -= next >> 2; lzo1x_decompress_safe() 155 next &= 3; lzo1x_decompress_safe() 176 next = get_unaligned_le16(ip); lzo1x_decompress_safe() 178 m_pos -= next >> 2; lzo1x_decompress_safe() 179 next &= 3; lzo1x_decompress_safe() 199 state = next; lzo1x_decompress_safe() 201 op += next; lzo1x_decompress_safe() 202 ip += next; lzo1x_decompress_safe() 225 state = next; lzo1x_decompress_safe() 226 t = next; lzo1x_decompress_safe()
|
/linux-4.4.14/arch/sh/mm/ |
H A D | tlb-sh5.c | 27 cpu_data->dtlb.next = cpu_data->dtlb.first; sh64_tlb_init() 38 cpu_data->itlb.next = cpu_data->itlb.first; sh64_tlb_init() 47 * sh64_next_free_dtlb_entry - Find the next available DTLB entry 51 return cpu_data->dtlb.next; sh64_next_free_dtlb_entry() 62 cpu_data->dtlb.next += cpu_data->dtlb.step; sh64_get_wired_dtlb_entry() 104 cpu_data->dtlb.next = entry; sh64_put_wired_dtlb_entry() 191 unsigned long long next; __update_tlb() local 214 next = tlbp->next; __update_tlb() 215 __flush_tlb_slot(next); __update_tlb() 218 : : "r" (next), "r" (pteh), "r" (ptel) ); __update_tlb() 220 next += TLB_STEP; __update_tlb() 221 if (next > tlbp->last) __update_tlb() 222 next = tlbp->first; __update_tlb() 223 tlbp->next = next; __update_tlb()
|
H A D | gup.c | 122 unsigned long next; gup_pmd_range() local 129 next = pmd_addr_end(addr, end); gup_pmd_range() 132 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) gup_pmd_range() 134 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 142 unsigned long next; gup_pud_range() local 149 next = pud_addr_end(addr, end); gup_pud_range() 152 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) gup_pud_range() 154 } while (pudp++, addr = next, addr != end); gup_pud_range() 168 unsigned long next; __get_user_pages_fast() local 190 next = pgd_addr_end(addr, end); __get_user_pages_fast() 193 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) __get_user_pages_fast() 195 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast() 222 unsigned long next; get_user_pages_fast() local 239 next = pgd_addr_end(addr, end); get_user_pages_fast() 242 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) get_user_pages_fast() 244 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
|
/linux-4.4.14/tools/include/linux/ |
H A D | list.h | 19 begin->prev->next = end->next; list_del_range() 20 end->next->prev = begin->prev; list_del_range() 29 for (; pos != (head); pos = pos->next)
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | switch_to.h | 6 struct task_struct *next); 15 "movl %P[task_canary](%[next]), %%ebx\n\t" \ 31 #define switch_to(prev, next, last) \ 66 : [next_sp] "m" (next->thread.sp), \ 67 [next_ip] "m" (next->thread.ip), \ 71 [next] "d" (next) \ 108 #define switch_to(prev, next, last) \ 111 "movq %P[threadrsp](%[next]),%%rsp\n\t" /* restore RSP */ \ 122 : [next] "S" (next), [prev] "D" (prev), \
|
H A D | mmu_context.h | 16 struct mm_struct *next) paravirt_activate_mm() 107 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 112 if (likely(prev != next)) { switch_mm() 115 this_cpu_write(cpu_tlbstate.active_mm, next); switch_mm() 117 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm() 124 * CPU 0: Write to a PTE for 'next' switch_mm() 126 * CPU 1: set bit 1 in next's mm_cpumask switch_mm() 140 * from next->pgd. TLB fills are special and can happen switch_mm() 147 load_cr3(next->pgd); switch_mm() 155 load_mm_cr4(next); switch_mm() 167 * exists. That means that next->context.ldt != switch_mm() 170 if (unlikely(prev->context.ldt != next->context.ldt)) switch_mm() 171 load_mm_ldt(next); switch_mm() 177 BUG_ON(this_cpu_read(cpu_tlbstate.active_mm) != next); switch_mm() 179 if (!cpumask_test_cpu(cpu, mm_cpumask(next))) { switch_mm() 186 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm() 196 load_cr3(next->pgd); switch_mm() 198 load_mm_cr4(next); switch_mm() 199 load_mm_ldt(next); switch_mm() 205 #define activate_mm(prev, next) \ 207 paravirt_activate_mm((prev), (next)); \ 208 switch_mm((prev), (next), NULL); \ 15 paravirt_activate_mm(struct mm_struct *prev, struct mm_struct *next) paravirt_activate_mm() argument
|
/linux-4.4.14/arch/arm/mm/ |
H A D | idmap.c | 25 unsigned long next; idmap_add_pmd() local 46 next = pmd_addr_end(addr, end); idmap_add_pmd() 49 } while (pmd++, addr = next, addr != end); idmap_add_pmd() 69 unsigned long next; idmap_add_pud() local 72 next = pud_addr_end(addr, end); idmap_add_pud() 73 idmap_add_pmd(pud, addr, next, prot); idmap_add_pud() 74 } while (pud++, addr = next, addr != end); idmap_add_pud() 81 unsigned long next; identity_mapping_add() local 94 next = pgd_addr_end(addr, end); identity_mapping_add() 95 idmap_add_pud(pgd, addr, next, prot); identity_mapping_add() 96 } while (pgd++, addr = next, addr != end); identity_mapping_add()
|
/linux-4.4.14/arch/cris/arch-v32/mm/ |
H A D | intmem.c | 110 struct intmem_allocation *next = crisv32_intmem_free() local 111 list_entry(allocation->entry.next, crisv32_intmem_free() 115 /* Join with prev and/or next if also free */ crisv32_intmem_free() 123 if ((next != &intmem_allocations) && crisv32_intmem_free() 124 (next->status == STATUS_FREE)) { crisv32_intmem_free() 125 allocation->size += next->size; crisv32_intmem_free() 126 list_del(&next->entry); crisv32_intmem_free() 127 kfree(next); crisv32_intmem_free()
|
/linux-4.4.14/arch/metag/include/asm/ |
H A D | mmu_context.h | 72 static inline void switch_mmu(struct mm_struct *prev, struct mm_struct *next) switch_mmu() argument 75 load_pgd(next->pgd, hard_processor_id()); switch_mmu() 87 next->pgd = prev->pgd; switch_mmu() 91 next->pgd[i] = ((pgd_t *) next->context.pgd_base)[i]; switch_mmu() 98 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 101 if (prev != next) switch_mm() 102 switch_mmu(prev, next); switch_mm()
|
/linux-4.4.14/drivers/usb/host/ |
H A D | ohci-dbg.c | 22 #define ohci_dbg_sw(ohci, next, size, format, arg...) \ 24 if (next != NULL) { \ 26 s_len = scnprintf (*next, *size, format, ## arg ); \ 27 *size -= s_len; *next += s_len; \ 32 /* Version for use where "next" is the address of a local variable */ 33 #define ohci_dbg_nosw(ohci, next, size, format, arg...) \ 36 s_len = scnprintf(*next, *size, format, ## arg); \ 37 *size -= s_len; *next += s_len; \ 45 char **next, ohci_dump_intr_mask() 48 ohci_dbg_sw (ohci, next, size, "%s 0x%08x%s%s%s%s%s%s%s%s%s\n", ohci_dump_intr_mask() 67 char **next, maybe_print_eds() 71 ohci_dbg_sw (ohci, next, size, "%s %08x\n", label, value); maybe_print_eds() 100 ohci_dump_status (struct ohci_hcd *controller, char **next, unsigned *size) ohci_dump_status() argument 106 ohci_dbg_sw (controller, next, size, ohci_dump_status() 113 ohci_dbg_sw (controller, next, size, ohci_dump_status() 128 ohci_dbg_sw (controller, next, size, ohci_dump_status() 139 next, size); ohci_dump_status() 142 next, size); ohci_dump_status() 147 next, size); ohci_dump_status() 151 next, size); ohci_dump_status() 154 next, size); ohci_dump_status() 158 next, size); ohci_dump_status() 161 next, size); ohci_dump_status() 164 ohci_readl (controller, ®s->donehead), next, size); ohci_dump_status() 167 #define dbg_port_sw(hc,num,value,next,size) \ 168 ohci_dbg_sw (hc, next, size, \ 193 char **next, ohci_dump_roothub() 203 ohci_dbg_sw (controller, next, size, ohci_dump_roothub() 214 ohci_dbg_sw (controller, next, size, ohci_dump_roothub() 221 ohci_dbg_sw (controller, next, size, ohci_dump_roothub() 235 dbg_port_sw (controller, i, temp, next, size); ohci_dump_roothub() 259 ohci_dbg (ohci, "%s td %p%s; urb %p index %d; hw next td %08x\n", ohci_dump_td() 317 ohci_dbg (ohci, "%s, ed %p state 0x%x type %s; next ed %08x\n", ohci_dump_ed() 491 char *next; fill_periodic_buffer() local 500 next = buf->page; fill_periodic_buffer() 503 temp = scnprintf (next, size, "size = %d\n", NUM_INTS); fill_periodic_buffer() 505 next += temp; fill_periodic_buffer() 514 temp = scnprintf (next, size, "%2d [%3d]:", i, ohci->load [i]); fill_periodic_buffer() 516 next += temp; fill_periodic_buffer() 519 temp = scnprintf (next, size, " ed%d/%p", fill_periodic_buffer() 522 next += temp; fill_periodic_buffer() 538 temp = scnprintf (next, size, fill_periodic_buffer() 554 next += temp; fill_periodic_buffer() 569 temp = scnprintf (next, size, "\n"); fill_periodic_buffer() 571 next += temp; fill_periodic_buffer() 587 char *next; fill_registers_buffer() local 593 next = buf->page; fill_registers_buffer() 600 ohci_dbg_nosw(ohci, &next, &size, fill_registers_buffer() 610 size -= scnprintf (next, size, fill_registers_buffer() 615 ohci_dump_status(ohci, &next, &size); fill_registers_buffer() 619 ohci_dbg_nosw(ohci, &next, &size, fill_registers_buffer() 624 temp = scnprintf (next, size, fill_registers_buffer() 629 next += temp; fill_registers_buffer() 632 temp = scnprintf (next, size, "fmremaining 0x%08x %sFR=0x%04x\n", fill_registers_buffer() 636 next += temp; fill_registers_buffer() 639 temp = scnprintf (next, size, "periodicstart 0x%04x\n", fill_registers_buffer() 642 next += temp; fill_registers_buffer() 645 temp = scnprintf (next, size, "lsthresh 0x%04x\n", fill_registers_buffer() 648 next += temp; fill_registers_buffer() 650 temp = scnprintf (next, size, "hub poll timer %s\n", fill_registers_buffer() 653 next += temp; fill_registers_buffer() 656 ohci_dump_roothub (ohci, 1, &next, &size); fill_registers_buffer() 41 ohci_dump_intr_mask( struct ohci_hcd *ohci, char *label, u32 mask, char **next, unsigned *size) ohci_dump_intr_mask() argument 63 maybe_print_eds( struct ohci_hcd *ohci, char *label, u32 value, char **next, unsigned *size) maybe_print_eds() argument 190 ohci_dump_roothub( struct ohci_hcd *controller, int verbose, char **next, unsigned *size) ohci_dump_roothub() argument
|
H A D | ehci-dbg.c | 135 ehci_dbg (ehci, "%s [%d] itd %p, next %08x, urb %p\n", dbg_itd() 166 ehci_dbg (ehci, "%s [%d] sitd %p, next %08x, urb %p\n", dbg_sitd() 419 char *next = *nextp; qh_lines() local 438 temp = scnprintf (next, size, qh_lines() 449 next += temp; qh_lines() 466 temp = snprintf (next, size, qh_lines() 481 next += temp; qh_lines() 486 temp = snprintf (next, size, "\n"); qh_lines() 490 next += temp; qh_lines() 494 *nextp = next; qh_lines() 503 char *next; fill_async_buffer() local 508 next = buf->output_buf; fill_async_buffer() 511 *next = 0; fill_async_buffer() 519 qh_lines (ehci, qh, &next, &size); fill_async_buffer() 521 temp = scnprintf(next, size, "\nunlink =\n"); fill_async_buffer() 523 next += temp; fill_async_buffer() 528 qh_lines(ehci, qh, &next, &size); fill_async_buffer() 542 char *next; fill_bandwidth_buffer() local 549 next = buf->output_buf; fill_bandwidth_buffer() 552 *next = 0; fill_bandwidth_buffer() 557 temp = scnprintf(next, size, fill_bandwidth_buffer() 560 next += temp; fill_bandwidth_buffer() 563 temp = scnprintf(next, size, fill_bandwidth_buffer() 568 next += temp; fill_bandwidth_buffer() 573 temp = scnprintf(next, size, fill_bandwidth_buffer() 578 next += temp; fill_bandwidth_buffer() 581 temp = scnprintf(next, size, fill_bandwidth_buffer() 586 next += temp; fill_bandwidth_buffer() 588 temp = scnprintf(next, size, fill_bandwidth_buffer() 591 next += temp; fill_bandwidth_buffer() 595 temp = scnprintf(next, size, fill_bandwidth_buffer() 600 next += temp; fill_bandwidth_buffer() 603 temp = scnprintf(next, size, fill_bandwidth_buffer() 611 next += temp; fill_bandwidth_buffer() 616 return next - buf->output_buf; fill_bandwidth_buffer() 627 char *next; fill_periodic_buffer() local 638 next = buf->output_buf; fill_periodic_buffer() 641 temp = scnprintf (next, size, "size = %d\n", ehci->periodic_size); fill_periodic_buffer() 643 next += temp; fill_periodic_buffer() 655 temp = scnprintf (next, size, "%4d: ", i); fill_periodic_buffer() 657 next += temp; fill_periodic_buffer() 665 temp = scnprintf (next, size, " qh%d-%04x/%p", fill_periodic_buffer() 673 next += temp; fill_periodic_buffer() 679 temp = scnprintf (next, size, fill_periodic_buffer() 682 next += temp; fill_periodic_buffer() 707 temp = scnprintf (next, size, fill_periodic_buffer() 726 temp = scnprintf (next, size, fill_periodic_buffer() 733 temp = scnprintf (next, size, fill_periodic_buffer() 739 temp = scnprintf (next, size, fill_periodic_buffer() 750 next += temp; fill_periodic_buffer() 753 temp = scnprintf (next, size, "\n"); fill_periodic_buffer() 755 next += temp; fill_periodic_buffer() 785 char *next, scratch [80]; fill_registers_buffer() local 791 next = buf->output_buf; fill_registers_buffer() 797 size = scnprintf (next, size, fill_registers_buffer() 809 temp = scnprintf (next, size, fill_registers_buffer() 818 next += temp; fill_registers_buffer() 834 temp = scnprintf (next, size, fill_registers_buffer() 839 next += temp; fill_registers_buffer() 843 temp = scnprintf (next, size, fill_registers_buffer() 846 next += temp; fill_registers_buffer() 861 temp = scnprintf (next, size, "structural params 0x%08x\n", i); fill_registers_buffer() 863 next += temp; fill_registers_buffer() 866 temp = scnprintf (next, size, "capability params 0x%08x\n", i); fill_registers_buffer() 868 next += temp; fill_registers_buffer() 873 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer() 875 next += temp; fill_registers_buffer() 879 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer() 881 next += temp; fill_registers_buffer() 885 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer() 887 next += temp; fill_registers_buffer() 889 temp = scnprintf (next, size, "uframe %04x\n", fill_registers_buffer() 892 next += temp; fill_registers_buffer() 898 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer() 900 next += temp; fill_registers_buffer() 902 temp = scnprintf (next, size, fill_registers_buffer() 907 next += temp; fill_registers_buffer() 912 temp = scnprintf(next, size, "async unlink qh %p\n", fill_registers_buffer() 916 next += temp; fill_registers_buffer() 920 temp = scnprintf (next, size, fill_registers_buffer() 925 next += temp; fill_registers_buffer() 927 temp = scnprintf (next, size, "complete %ld unlink %ld\n", fill_registers_buffer() 930 next += temp; fill_registers_buffer()
|
H A D | xhci-ext-caps.h | 94 * Return the next extended capability pointer register. 100 * in XHCI_HCC_PARAMS_OFFSET. If searching for the next extended capability, 103 * Returns 0 if there is no next extended capability register or returns the register offset 108 u32 next; xhci_find_next_cap_offset() local 110 next = readl(base + ext_offset); xhci_find_next_cap_offset() 114 next = XHCI_HCC_EXT_CAPS(next); xhci_find_next_cap_offset() 117 /* Find the next extended capability */ xhci_find_next_cap_offset() 118 next = XHCI_EXT_CAPS_NEXT(next); xhci_find_next_cap_offset() 121 if (!next) xhci_find_next_cap_offset() 127 return ext_offset + (next << 2); xhci_find_next_cap_offset()
|
H A D | fhci-q.c | 79 td = list_entry(ed->td_list.next, struct td, node); peek_td_from_ed() 91 td = list_entry(frame->tds_list.next, struct td, frame_lh); fhci_remove_td_from_frame() 92 list_del_init(frame->tds_list.next); fhci_remove_td_from_frame() 104 td = list_entry(frame->tds_list.next, struct td, frame_lh); fhci_peek_td_from_frame() 116 td = list_entry(ed->td_list.next, struct td, node); fhci_remove_td_from_ed() 117 list_del_init(ed->td_list.next); fhci_remove_td_from_ed() 119 /* if this TD was the ED's head, find next TD */ fhci_remove_td_from_ed() 121 ed->td_head = list_entry(ed->td_list.next, struct td, fhci_remove_td_from_ed() 136 td = list_entry(p_list->done_list.next, struct td, node); fhci_remove_td_from_done_list() 137 list_del_init(p_list->done_list.next); fhci_remove_td_from_done_list() 151 /* If this TD was the ED's head,find next TD */ fhci_move_td_from_ed_to_done_list() 153 ed->td_head = list_entry(ed->td_list.next, struct td, node); fhci_move_td_from_ed_to_done_list() 176 /* if this TD was the ED's head,find the next TD */ free_urb_priv() 178 ed->td_head = list_entry(ed->td_list.next, struct td, node); free_urb_priv() 186 /* if this TD was the ED's head,find next TD */ free_urb_priv()
|
/linux-4.4.14/scripts/gdb/linux/ |
H A D | lists.py | 35 n = c['next'].dereference() 37 if p['next'] != c.address: 38 gdb.write('prev.next != current: ' 56 gdb.write('next.prev != current: ' 58 'next@{n_addr}={n}\n'.format( 66 gdb.write('next is not accessible: '
|
H A D | modules.py | 26 entry = modules['next'] 31 entry = entry['next'] 84 entry = source_list['next'] 92 entry = entry['next']
|
/linux-4.4.14/arch/mn10300/include/asm/ |
H A D | switch_to.h | 22 #define switch_fpu(prev, next) \ 31 #define switch_fpu(prev, next) do {} while (0) 37 struct thread_struct *next, 40 #define switch_to(prev, next, last) \ 42 switch_fpu(prev, next); \ 44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
H A D | mmu_context.h | 140 * @next: The incoming MM context. 143 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 148 if (prev != next) { switch_mm() 150 per_cpu(cpu_tlbstate, cpu).active_mm = next; switch_mm() 152 cpu_ran_vm(cpu, next); switch_mm() 153 PTBR = (unsigned long) next->pgd; switch_mm() 154 activate_context(next); switch_mm() 159 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.4.14/sound/core/seq/ |
H A D | seq_prioq.c | 38 * |next 168 f->tail->next = cell; snd_seq_prioq_cell_in() 170 cell->next = NULL; snd_seq_prioq_cell_in() 193 /* move cursor to next cell */ snd_seq_prioq_cell_in() 195 cur = cur->next; snd_seq_prioq_cell_in() 205 prev->next = cell; snd_seq_prioq_cell_in() 206 cell->next = cur; snd_seq_prioq_cell_in() 231 f->head = cell->next; snd_seq_prioq_cell_out() 237 cell->next = NULL; snd_seq_prioq_cell_out() 292 register struct snd_seq_event_cell *cell, *next; snd_seq_prioq_leave() local 301 next = cell->next; snd_seq_prioq_leave() 305 f->head = cell->next; snd_seq_prioq_leave() 307 prev->next = cell->next; snd_seq_prioq_leave() 310 f->tail = cell->next; snd_seq_prioq_leave() 313 cell->next = NULL; snd_seq_prioq_leave() 317 freeprev->next = cell; snd_seq_prioq_leave() 331 cell = next; snd_seq_prioq_leave() 337 freenext = freefirst->next; snd_seq_prioq_leave() 402 struct snd_seq_event_cell *cell, *next; snd_seq_prioq_remove_events() local 412 next = cell->next; snd_seq_prioq_remove_events() 418 f->head = cell->next; snd_seq_prioq_remove_events() 420 prev->next = cell->next; snd_seq_prioq_remove_events() 424 f->tail = cell->next; snd_seq_prioq_remove_events() 428 cell->next = NULL; snd_seq_prioq_remove_events() 432 freeprev->next = cell; snd_seq_prioq_remove_events() 439 cell = next; snd_seq_prioq_remove_events() 445 freenext = freefirst->next; snd_seq_prioq_remove_events()
|
H A D | seq_fifo.c | 134 f->tail->next = cell; snd_seq_fifo_event_in() 157 f->head = cell->next; fifo_cell_out() 163 cell->next = NULL; fifo_cell_out() 215 cell->next = f->head; snd_seq_fifo_cell_putback() 236 struct snd_seq_event_cell *cell, *next, *oldhead; snd_seq_fifo_resize() local 263 for (cell = oldhead; cell; cell = next) { snd_seq_fifo_resize() 264 next = cell->next; snd_seq_fifo_resize()
|
/linux-4.4.14/arch/mn10300/kernel/ |
H A D | kprobes.c | 125 * Calculate what the PC will be after executing next instruction 133 u8 opc, *pc, *sp, *next; find_nextpc() local 135 next = 0; find_nextpc() 144 next = pc + size; find_nextpc() 152 next = pc + x8; find_nextpc() 160 next = pc + x16; find_nextpc() 168 next = pc + x32; find_nextpc() 174 next = (u8 *)regs->mdr; find_nextpc() 183 next = (u8 *)x32; find_nextpc() 189 next = pc + 2; find_nextpc() 195 next = (u8 *)regs->a0; find_nextpc() 198 next = (u8 *)regs->a1; find_nextpc() 201 next = (u8 *)regs->a2; find_nextpc() 204 next = (u8 *)regs->a3; find_nextpc() 212 next = (u8 *)x32; find_nextpc() 218 next = (u8 *)x32; find_nextpc() 226 next = pc + 3; find_nextpc() 233 next = pc + x8; find_nextpc() 242 next = pc + x16; find_nextpc() 244 next = pc + 4; find_nextpc() 254 next = pc + x32; find_nextpc() 258 /* SETLB - loads the next four bytes into the LIR reg */ find_nextpc() 265 return (unsigned)next; find_nextpc() 274 u8 opc, *pc, *sp, *next; singlestep_branch_setup() local 276 next = NULL; singlestep_branch_setup() 286 next = pc; singlestep_branch_setup() 292 next = pc + 5; singlestep_branch_setup() 300 next = pc + 7; singlestep_branch_setup() 304 next = pc + 3; singlestep_branch_setup() 305 regs->mdr = (unsigned) next; singlestep_branch_setup() 310 next = pc + 3; singlestep_branch_setup() 311 *(unsigned *)sp = (unsigned) next; singlestep_branch_setup() 315 next = pc + 2; singlestep_branch_setup() 324 next = pc + 4; singlestep_branch_setup() 327 next = pc; singlestep_branch_setup() 330 next = pc + 2; singlestep_branch_setup() 331 *(unsigned *) sp = (unsigned) next; singlestep_branch_setup() 334 next = pc + 2; singlestep_branch_setup() 335 *(unsigned *)(sp + 4) = (unsigned) next; singlestep_branch_setup() 342 next = pc + 4; singlestep_branch_setup() 350 next = pc + 6; singlestep_branch_setup() 358 return (unsigned) next; singlestep_branch_setup()
|
/linux-4.4.14/arch/arm64/mm/ |
H A D | kasan_init.c | 30 unsigned long next; kasan_early_pte_populate() local 37 next = addr + PAGE_SIZE; kasan_early_pte_populate() 40 } while (pte++, addr = next, addr != end && pte_none(*pte)); kasan_early_pte_populate() 48 unsigned long next; kasan_early_pmd_populate() local 55 next = pmd_addr_end(addr, end); kasan_early_pmd_populate() 56 kasan_early_pte_populate(pmd, addr, next); kasan_early_pmd_populate() 57 } while (pmd++, addr = next, addr != end && pmd_none(*pmd)); kasan_early_pmd_populate() 65 unsigned long next; kasan_early_pud_populate() local 72 next = pud_addr_end(addr, end); kasan_early_pud_populate() 73 kasan_early_pmd_populate(pud, addr, next); kasan_early_pud_populate() 74 } while (pud++, addr = next, addr != end && pud_none(*pud)); kasan_early_pud_populate() 81 unsigned long next; kasan_map_early_shadow() local 86 next = pgd_addr_end(addr, end); kasan_map_early_shadow() 87 kasan_early_pud_populate(pgd, addr, next); kasan_map_early_shadow() 88 } while (pgd++, addr = next, addr != end); kasan_map_early_shadow()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | mm.c | 54 struct nvkm_mm_node *next = node(this, next); nvkm_mm_free() local 62 if (next && next->type == NVKM_MM_TYPE_NONE) { nvkm_mm_free() 63 next->offset = this->offset; nvkm_mm_free() 64 next->length += this->length; nvkm_mm_free() 114 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_head() local 133 next = node(this, next); nvkm_mm_head() 134 if (next && next->type != type) nvkm_mm_head() 188 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_tail() local 206 next = node(this, next); nvkm_mm_tail() 207 if (next && next->type != type) { nvkm_mm_tail() 209 c = next->offset - e; nvkm_mm_tail() 241 u32 next; nvkm_mm_init() local 245 next = prev->offset + prev->length; nvkm_mm_init() 246 if (next != offset) { nvkm_mm_init() 247 BUG_ON(next > offset); nvkm_mm_init() 251 node->offset = next; nvkm_mm_init() 252 node->length = offset - next; nvkm_mm_init()
|
/linux-4.4.14/arch/nios2/mm/ |
H A D | mmu_context.c | 60 /* Return the next pid */ get_new_context() 80 void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 89 if (unlikely(CTX_VERSION(next->context) != switch_mm() 91 next->context = get_new_context(); switch_mm() 94 pgd_current = next->pgd; switch_mm() 97 set_context(next->context); switch_mm() 106 void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 108 next->context = get_new_context(); activate_mm() 109 set_context(next->context); activate_mm() 110 pgd_current = next->pgd; activate_mm()
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | mmu_context.h | 37 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next); 51 extern void switch_cop(struct mm_struct *next); mmu_context_init() 59 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 63 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); switch_mm() 67 tsk->thread.pgdir = next->pgd; switch_mm() 72 get_paca()->pgd = next->pgd; switch_mm() 75 if (prev == next) switch_mm() 79 /* Switch coprocessor context only if prev or next uses a coprocessor */ switch_mm() 80 if (prev->context.acop || next->context.acop) switch_mm() 81 switch_cop(next); switch_mm() 96 switch_slb(tsk, next); switch_mm() 99 switch_mmu_context(prev, next); switch_mm() 110 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 115 switch_mm(prev, next, current); activate_mm()
|
H A D | bootx.h | 28 u32 next; member in struct:bootx_dt_prop 44 u32 next; member in struct:bootx_dt_node
|
/linux-4.4.14/drivers/parisc/ |
H A D | hppb.c | 29 struct hppb_card *next; member in struct:hppb_card 34 .next = NULL, 53 while(card->next) { hppb_probe() 54 card = card->next; hppb_probe() 58 card->next = kzalloc(sizeof(struct hppb_card), GFP_KERNEL); hppb_probe() 59 if(!card->next) { hppb_probe() 63 card = card->next; hppb_probe()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/include/nvif/ |
H A D | list.h | 12 * The above copyright notice and this permission notice (including the next 92 * struct foo *iterator, *next; 93 * list_for_each_entry_safe(iterator, next, &bar.list_of_foos, entry) { 111 struct list_head *next, *prev; member in struct:list_head 130 list->next = list->prev = list; INIT_LIST_HEAD() 135 struct list_head *prev, struct list_head *next) __list_add() 137 next->prev = entry; __list_add() 138 entry->next = next; __list_add() 140 prev->next = entry; __list_add() 161 __list_add(entry, head, head->next); list_add() 186 __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 188 next->prev = prev; __list_del() 189 prev->next = next; __list_del() 209 __list_del(entry->prev, entry->next); list_del() 215 __list_del(entry->prev, entry->next); list_del_init() 222 __list_del(list->prev, list->next); list_move_tail() 237 return head->next == head; list_empty() 277 list_entry((ptr)->next, type, member) 315 for (pos = __container_of((head)->next, pos, member); \ 317 pos = __container_of(pos->member.next, pos, member)) 327 for (pos = __container_of((head)->next, pos, member), \ 328 tmp = __container_of(pos->member.next, pos, member); \ 330 pos = tmp, tmp = __container_of(pos->member.next, tmp, member)) 339 for (pos = __container_of(pos->member.next, pos, member); \ 341 pos = __container_of(pos->member.next, pos, member)) 351 pos = __container_of(pos->member.next, pos, member)) 134 __list_add(struct list_head *entry, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.4.14/arch/tile/include/asm/ |
H A D | switch_to.h | 36 #define prepare_arch_switch(next) _prepare_arch_switch(next) 37 void _prepare_arch_switch(struct task_struct *next); 40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next))) 42 struct task_struct *next); 46 struct task_struct *next,
|
H A D | cacheflush.h | 81 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); __flush_buffer() local 83 while (next < finish) { __flush_buffer() 84 __insn_flush(next); __flush_buffer() 85 next += CHIP_FLUSH_STRIDE(); __flush_buffer() 92 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); __finv_buffer() local 94 while (next < finish) { __finv_buffer() 95 __insn_finv(next); __finv_buffer() 96 next += CHIP_FINV_STRIDE(); __finv_buffer() 125 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); __inv_buffer() local 127 while (next < finish) { __inv_buffer() 128 __insn_inv(next); __inv_buffer() 129 next += CHIP_INV_STRIDE(); __inv_buffer()
|
/linux-4.4.14/arch/hexagon/include/asm/ |
H A D | mmu_context.h | 66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 75 if (next->context.generation < prev->context.generation) { switch_mm() 77 next->pgd[l1] = init_mm.pgd[l1]; switch_mm() 79 next->context.generation = prev->context.generation; switch_mm() 82 __vmnewmap((void *)next->context.ptbase); switch_mm() 88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 93 switch_mm(prev, next, current_thread_info()->task); activate_mm()
|
/linux-4.4.14/drivers/pci/hotplug/ |
H A D | cpqphp_ctrl.c | 135 slot = slot->next; cpqhp_find_slot() 321 if (!((*head)->next)) sort_by_size() 328 if (((*head)->next) && sort_by_size() 329 ((*head)->length > (*head)->next->length)) { sort_by_size() 332 *head = (*head)->next; sort_by_size() 333 current_res->next = (*head)->next; sort_by_size() 334 (*head)->next = current_res; sort_by_size() 339 while (current_res->next && current_res->next->next) { sort_by_size() 340 if (current_res->next->length > current_res->next->next->length) { sort_by_size() 342 next_res = current_res->next; sort_by_size() 343 current_res->next = current_res->next->next; sort_by_size() 344 current_res = current_res->next; sort_by_size() 345 next_res->next = current_res->next; sort_by_size() 346 current_res->next = next_res; sort_by_size() 348 current_res = current_res->next; sort_by_size() 369 if (!((*head)->next)) sort_by_max_size() 376 if (((*head)->next) && sort_by_max_size() 377 ((*head)->length < (*head)->next->length)) { sort_by_max_size() 380 *head = (*head)->next; sort_by_max_size() 381 current_res->next = (*head)->next; sort_by_max_size() 382 (*head)->next = current_res; sort_by_max_size() 387 while (current_res->next && current_res->next->next) { sort_by_max_size() 388 if (current_res->next->length < current_res->next->next->length) { sort_by_max_size() 390 next_res = current_res->next; sort_by_max_size() 391 current_res->next = current_res->next->next; sort_by_max_size() 392 current_res = current_res->next; sort_by_max_size() 393 next_res->next = current_res->next; sort_by_max_size() 394 current_res->next = next_res; sort_by_max_size() 396 current_res = current_res->next; sort_by_max_size() 460 split_node->next = node; do_pre_bridge_resource_split() 468 *head = node->next; do_pre_bridge_resource_split() 471 while (prevnode->next != node) do_pre_bridge_resource_split() 472 prevnode = prevnode->next; do_pre_bridge_resource_split() 474 prevnode->next = node->next; do_pre_bridge_resource_split() 476 node->next = NULL; do_pre_bridge_resource_split() 501 while (node->next) { do_bridge_resource_split() 503 node = node->next; do_bridge_resource_split() 556 for (node = *head; node; node = node->next) { get_io_resource() 581 split_node->next = node->next; get_io_resource() 582 node->next = split_node; get_io_resource() 600 split_node->next = node->next; get_io_resource() 601 node->next = split_node; get_io_resource() 612 *head = node->next; get_io_resource() 615 while (prevnode->next != node) get_io_resource() 616 prevnode = prevnode->next; get_io_resource() 618 prevnode->next = node->next; get_io_resource() 620 node->next = NULL; get_io_resource() 650 for (max = *head; max; max = max->next) { get_max_resource() 652 * instead we'll continue to the next. get_max_resource() 677 split_node->next = max->next; get_max_resource() 678 max->next = split_node; get_max_resource() 695 split_node->next = max->next; get_max_resource() 696 max->next = split_node; get_max_resource() 706 *head = max->next; get_max_resource() 708 while (temp && temp->next != max) get_max_resource() 709 temp = temp->next; get_max_resource() 712 temp->next = max->next; get_max_resource() 715 max->next = NULL; get_max_resource() 747 for (node = *head; node; node = node->next) { get_resource() 774 split_node->next = node->next; get_resource() 775 node->next = split_node; get_resource() 794 split_node->next = node->next; get_resource() 795 node->next = split_node; get_resource() 802 *head = node->next; get_resource() 805 while (prevnode->next != node) get_resource() 806 prevnode = prevnode->next; get_resource() 808 prevnode->next = node->next; get_resource() 810 node->next = NULL; get_resource() 838 dbg("*head->next = %p\n",(*head)->next); cpqhp_resource_sort_and_combine() 840 if (!(*head)->next) cpqhp_resource_sort_and_combine() 844 dbg("*head->next->base = 0x%x\n",(*head)->next->base); cpqhp_resource_sort_and_combine() 849 if (((*head)->next) && cpqhp_resource_sort_and_combine() 850 ((*head)->base > (*head)->next->base)) { cpqhp_resource_sort_and_combine() 852 (*head) = (*head)->next; cpqhp_resource_sort_and_combine() 853 node1->next = (*head)->next; cpqhp_resource_sort_and_combine() 854 (*head)->next = node1; cpqhp_resource_sort_and_combine() 860 while (node1->next && node1->next->next) { cpqhp_resource_sort_and_combine() 861 if (node1->next->base > node1->next->next->base) { cpqhp_resource_sort_and_combine() 863 node2 = node1->next; cpqhp_resource_sort_and_combine() 864 node1->next = node1->next->next; cpqhp_resource_sort_and_combine() 865 node1 = node1->next; cpqhp_resource_sort_and_combine() 866 node2->next = node1->next; cpqhp_resource_sort_and_combine() 867 node1->next = node2; cpqhp_resource_sort_and_combine() 869 node1 = node1->next; cpqhp_resource_sort_and_combine() 875 while (node1 && node1->next) { cpqhp_resource_sort_and_combine() 876 if ((node1->base + node1->length) == node1->next->base) { cpqhp_resource_sort_and_combine() 879 node1->length += node1->next->length; cpqhp_resource_sort_and_combine() 880 node2 = node1->next; cpqhp_resource_sort_and_combine() 881 node1->next = node1->next->next; cpqhp_resource_sort_and_combine() 884 node1 = node1->next; cpqhp_resource_sort_and_combine() 971 struct pci_func *next; cpqhp_slot_create() local 977 new_slot->next = NULL; cpqhp_slot_create() 983 next = cpqhp_slot_list[busnumber]; cpqhp_slot_create() 984 while (next->next != NULL) cpqhp_slot_create() 985 next = next->next; cpqhp_slot_create() 986 next->next = new_slot; cpqhp_slot_create() 1000 struct pci_func *next; slot_remove() local 1005 next = cpqhp_slot_list[old_slot->bus]; slot_remove() 1006 if (next == NULL) slot_remove() 1009 if (next == old_slot) { slot_remove() 1010 cpqhp_slot_list[old_slot->bus] = old_slot->next; slot_remove() 1016 while ((next->next != old_slot) && (next->next != NULL)) slot_remove() 1017 next = next->next; slot_remove() 1019 if (next->next == old_slot) { slot_remove() 1020 next->next = old_slot->next; slot_remove() 1039 struct pci_func *next; bridge_slot_remove() local 1045 next = cpqhp_slot_list[tempBus]; bridge_slot_remove() 1047 while (!slot_remove(next)) bridge_slot_remove() 1048 next = cpqhp_slot_list[tempBus]; bridge_slot_remove() 1051 next = cpqhp_slot_list[bridge->bus]; bridge_slot_remove() 1053 if (next == NULL) bridge_slot_remove() 1056 if (next == bridge) { bridge_slot_remove() 1057 cpqhp_slot_list[bridge->bus] = bridge->next; bridge_slot_remove() 1061 while ((next->next != bridge) && (next->next != NULL)) bridge_slot_remove() 1062 next = next->next; bridge_slot_remove() 1064 if (next->next != bridge) bridge_slot_remove() 1066 next->next = bridge->next; bridge_slot_remove() 1094 while (func->next != NULL) { cpqhp_slot_find() 1095 func = func->next; cpqhp_slot_find() 1144 for (slot = ctrl->slot; slot; slot = slot->next) { set_controller_speed() 1580 /* next, we will instantiate the linux pci_dev structures (with board_added() 1758 for (ctrl = cpqhp_ctrl_list; ctrl; ctrl=ctrl->next) event_thread() 2171 /* Get ready for next iteration */ switch_leds() 2222 /* Get ready for next iteration */ cpqhp_hardware_test() 2232 /* Get ready for next iteration */ cpqhp_hardware_test() 2316 /* The following loop skips to the next present function configure_new_device() 2459 dbg("(base, len, next) (%x, %x, %p)\n", io_node->base, configure_new_function() 2460 io_node->length, io_node->next); configure_new_function() 2462 dbg("(base, len, next) (%x, %x, %p)\n", mem_node->base, configure_new_function() 2463 mem_node->length, mem_node->next); configure_new_function() 2465 dbg("(base, len, next) (%x, %x, %p)\n", p_mem_node->base, configure_new_function() 2466 p_mem_node->length, p_mem_node->next); configure_new_function() 2514 bus_node->next = NULL; configure_new_function() 2519 io_node->next = NULL; configure_new_function() 2532 mem_node->next = NULL; configure_new_function() 2542 p_mem_node->next = NULL; configure_new_function() 2612 hold_bus_node->next = func->bus_head; configure_new_function() 2654 hold_IO_node->next = func->io_head; configure_new_function() 2671 hold_IO_node->next = func->io_head; configure_new_function() 2676 hold_IO_node->next = func->io_head; configure_new_function() 2704 hold_mem_node->next = func->mem_head; configure_new_function() 2723 hold_mem_node->next = func->mem_head; configure_new_function() 2728 hold_mem_node->next = func->mem_head; configure_new_function() 2757 hold_p_mem_node->next = func->p_mem_head; configure_new_function() 2774 hold_p_mem_node->next = func->p_mem_head; configure_new_function() 2779 hold_p_mem_node->next = func->p_mem_head; configure_new_function() 2828 dbg("Got io_node start = %8.8x, length = %8.8x next (%p)\n", configure_new_function() 2829 io_node->base, io_node->length, io_node->next); configure_new_function() 2836 io_node->next = func->io_head; configure_new_function() 2852 p_mem_node->next = func->p_mem_head; configure_new_function() 2868 mem_node->next = func->mem_head; configure_new_function()
|
H A D | cpqphp_pci.c | 433 /* this loop skips to the next present function cpqhp_save_config() 523 /* this loop skips to the next present function cpqhp_save_slot_config() 562 struct pci_func *next; cpqhp_save_base_addr_length() local 581 next = cpqhp_slot_list[sub_bus]; cpqhp_save_base_addr_length() 583 while (next != NULL) { cpqhp_save_base_addr_length() 584 rc = cpqhp_save_base_addr_length(ctrl, next); cpqhp_save_base_addr_length() 588 next = next->next; cpqhp_save_base_addr_length() 672 /* find the next device in this slot */ cpqhp_save_base_addr_length() 742 bus_node->next = func->bus_head; cpqhp_save_used_resources() 757 io_node->next = func->io_head; cpqhp_save_used_resources() 773 mem_node->next = func->mem_head; cpqhp_save_used_resources() 789 p_mem_node->next = func->p_mem_head; cpqhp_save_used_resources() 822 io_node->next = func->io_head; cpqhp_save_used_resources() 839 p_mem_node->next = func->p_mem_head; cpqhp_save_used_resources() 856 mem_node->next = func->mem_head; cpqhp_save_used_resources() 893 io_node->next = func->io_head; cpqhp_save_used_resources() 910 p_mem_node->next = func->p_mem_head; cpqhp_save_used_resources() 927 mem_node->next = func->mem_head; cpqhp_save_used_resources() 935 /* find the next device in this slot */ cpqhp_save_used_resources() 958 struct pci_func *next; cpqhp_configure_board() local 985 next = cpqhp_slot_list[sub_bus]; cpqhp_configure_board() 987 while (next != NULL) { cpqhp_configure_board() 988 rc = cpqhp_configure_board(ctrl, next); cpqhp_configure_board() 992 next = next->next; cpqhp_configure_board() 1039 struct pci_func *next; cpqhp_valid_replace() local 1083 next = cpqhp_slot_list[secondary_bus]; cpqhp_valid_replace() 1085 while (next != NULL) { cpqhp_valid_replace() 1086 rc = cpqhp_valid_replace(ctrl, next); cpqhp_valid_replace() 1090 next = next->next; cpqhp_valid_replace() 1154 /* Get the next function */ cpqhp_valid_replace() 1318 io_node->next = ctrl->io_head; cpqhp_find_available_resources() 1321 io_node->next = func->io_head; cpqhp_find_available_resources() 1341 mem_node->next = ctrl->mem_head; cpqhp_find_available_resources() 1344 mem_node->next = func->mem_head; cpqhp_find_available_resources() 1366 p_mem_node->next = ctrl->p_mem_head; cpqhp_find_available_resources() 1369 p_mem_node->next = func->p_mem_head; cpqhp_find_available_resources() 1389 bus_node->next = ctrl->bus_head; cpqhp_find_available_resources() 1392 bus_node->next = func->bus_head; cpqhp_find_available_resources() 1435 t_node = node->next; cpqhp_return_board_resources() 1443 t_node = node->next; cpqhp_return_board_resources() 1451 t_node = node->next; cpqhp_return_board_resources() 1459 t_node = node->next; cpqhp_return_board_resources() 1487 res = res->next; cpqhp_destroy_resource_list() 1496 res = res->next; cpqhp_destroy_resource_list() 1505 res = res->next; cpqhp_destroy_resource_list() 1514 res = res->next; cpqhp_destroy_resource_list() 1534 res = res->next; cpqhp_destroy_board_resources() 1543 res = res->next; cpqhp_destroy_board_resources() 1552 res = res->next; cpqhp_destroy_board_resources() 1561 res = res->next; cpqhp_destroy_board_resources()
|
H A D | ibmphp_res.c | 417 range_cur = range_cur->next; add_bus_range() 433 range->next = range_cur; add_bus_range() 438 range->next = NULL; add_bus_range() 439 range_prev->next = range; add_bus_range() 444 range_prev->next = range; add_bus_range() 445 range->next = range_cur; add_bus_range() 452 range_cur = range_cur->next; add_bus_range() 489 if (res->next) update_resources() 490 res = res->next; update_resources() 503 res = res->next; update_resources() 543 range = range->next; fix_me() 546 if (res->next) fix_me() 547 res = res->next; fix_me() 591 * the ptrs to next and nextRange if needed. 643 range_cur = range_cur->next; ibmphp_add_resource() 680 res->next = NULL; ibmphp_add_resource() 692 if (res_cur->next) ibmphp_add_resource() 693 res_cur = res_cur->next; ibmphp_add_resource() 702 res->next = NULL; ibmphp_add_resource() 710 res_cur = res_cur->next; ibmphp_add_resource() 714 res_prev->next = res; ibmphp_add_resource() 715 res->next = NULL; ibmphp_add_resource() 733 res_prev->next = res; ibmphp_add_resource() 737 res->next = res_cur; ibmphp_add_resource() 744 res->next = NULL; ibmphp_add_resource() 762 res->next = NULL; ibmphp_add_resource() 822 if (res_cur->next) ibmphp_remove_resource() 823 res_cur = res_cur->next; ibmphp_remove_resource() 845 if (mem_cur->next) ibmphp_remove_resource() 846 mem_cur = mem_cur->next; ibmphp_remove_resource() 857 bus_cur->firstPFMemFromMem = res_cur->next; ibmphp_remove_resource() 859 res_prev->next = res_cur->next; ibmphp_remove_resource() 864 if (res_cur->next) ibmphp_remove_resource() 865 res_cur = res_cur->next; ibmphp_remove_resource() 880 if (res_cur->next) { ibmphp_remove_resource() 883 bus_cur->firstIO = res_cur->next; ibmphp_remove_resource() 886 bus_cur->firstMem = res_cur->next; ibmphp_remove_resource() 889 bus_cur->firstPFMem = res_cur->next; ibmphp_remove_resource() 920 if (res_cur->next) { ibmphp_remove_resource() 922 res_prev->next = res_cur->next; ibmphp_remove_resource() 924 res_prev->nextRange = res_cur->next; ibmphp_remove_resource() 926 res_prev->next = NULL; ibmphp_remove_resource() 929 res_prev->next = NULL; ibmphp_remove_resource() 960 range = range->next; find_range() 1084 if (!res_cur->next) { ibmphp_check_resource() 1203 if (res_cur->next) ibmphp_check_resource() 1204 res_cur = res_cur->next; ibmphp_check_resource() 1260 range = range->next; ibmphp_check_resource() 1326 range = range->next; ibmphp_check_resource() 1386 if (res_cur->next) ibmphp_remove_bus() 1387 res_cur = res_cur->next; ibmphp_remove_bus() 1399 if (res_cur->next) ibmphp_remove_bus() 1400 res_cur = res_cur->next; ibmphp_remove_bus() 1412 if (res_cur->next) ibmphp_remove_bus() 1413 res_cur = res_cur->next; ibmphp_remove_bus() 1426 res_cur = res_cur->next; ibmphp_remove_bus() 1460 range_cur = range_cur->next; remove_ranges() 1474 range_cur = range_cur->next; remove_ranges() 1488 range_cur = range_cur->next; remove_ranges() 1534 if (res_cur->next) ibmphp_find_resource() 1535 res_cur = res_cur->next; ibmphp_find_resource() 1548 res_cur = res_cur->next; ibmphp_find_resource() 1581 struct list_head *next; ibmphp_free_resources() local 1585 list_for_each_safe (tmp, next, &gbuses) { ibmphp_free_resources() 1593 range_cur = range_cur->next; ibmphp_free_resources() 1604 range_cur = range_cur->next; ibmphp_free_resources() 1615 range_cur = range_cur->next; ibmphp_free_resources() 1625 if (res_cur->next) ibmphp_free_resources() 1626 res_cur = res_cur->next; ibmphp_free_resources() 1638 if (res_cur->next) ibmphp_free_resources() 1639 res_cur = res_cur->next; ibmphp_free_resources() 1651 if (res_cur->next) ibmphp_free_resources() 1652 res_cur = res_cur->next; ibmphp_free_resources() 1665 res_cur = res_cur->next; ibmphp_free_resources() 1699 for (pfmem_cur = bus_cur->firstPFMem, pfmem_prev = NULL; pfmem_cur; pfmem_prev = pfmem_cur, pfmem_cur = pfmem_cur->next) { once_over() 1702 pfmem_prev->next = pfmem_cur->next; once_over() 1704 bus_cur->firstPFMem = pfmem_cur->next; once_over() 1707 pfmem_cur->next = NULL; once_over() 1713 pfmem_cur->next = bus_cur->firstPFMemFromMem; once_over() 1747 pfmem->next = bus_cur->firstPFMemFromMem; ibmphp_add_pfmem_from_mem() 1749 pfmem->next = NULL; ibmphp_add_pfmem_from_mem() 1812 range = range->next; ibmphp_print_test() 1822 range = range->next; ibmphp_print_test() 1833 range = range->next; ibmphp_print_test() 1846 if (res->next) ibmphp_print_test() 1847 res = res->next; ibmphp_print_test() 1861 if (res->next) ibmphp_print_test() 1862 res = res->next; ibmphp_print_test() 1876 if (res->next) ibmphp_print_test() 1877 res = res->next; ibmphp_print_test() 1892 res = res->next; ibmphp_print_test() 1920 range_cur = range_cur->next; range_exists_already()
|
/linux-4.4.14/drivers/staging/lustre/lustre/lov/ |
H A D | lovsub_dev.c | 94 const char *name, struct lu_device *next) lovsub_device_init() 100 next->ld_site = d->ld_site; lovsub_device_init() 101 ldt = next->ld_type; lovsub_device_init() 103 rc = ldt->ldt_ops->ldto_device_init(env, next, ldt->ldt_name, NULL); lovsub_device_init() 105 next->ld_site = NULL; lovsub_device_init() 109 lu_device_get(next); lovsub_device_init() 110 lu_ref_add(&next->ld_reference, "lu-stack", &lu_site_init); lovsub_device_init() 111 lsd->acid_next = lu2cl_dev(next); lovsub_device_init() 118 struct lu_device *next; lovsub_device_fini() local 122 next = cl2lu_dev(lsd->acid_next); lovsub_device_fini() 125 return next; lovsub_device_fini() 132 struct lu_device *next = cl2lu_dev(lsd->acid_next); lovsub_device_free() local 140 return next; lovsub_device_free() 93 lovsub_device_init(const struct lu_env *env, struct lu_device *d, const char *name, struct lu_device *next) lovsub_device_init() argument
|
/linux-4.4.14/drivers/scsi/arm/ |
H A D | msgqueue.c | 30 msgq->free = mq->next; mqe_alloc() 44 mq->next = msgq->free; mqe_free() 62 msgq->entries[i].next = &msgq->entries[i + 1]; msgqueue_initialise() 64 msgq->entries[NR_MESSAGES - 1].next = NULL; msgqueue_initialise() 88 for (mq = msgq->qe; mq; mq = mq->next) msgqueue_msglength() 105 for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); msgqueue_getmsg() 134 mq->next = NULL; msgqueue_addmsg() 138 mqp = &(*mqp)->next; msgqueue_addmsg() 156 mqnext = mq->next; msgqueue_flush()
|
/linux-4.4.14/drivers/media/pci/zoran/ |
H A D | videocodec.c | 60 struct attached_list *next; member in struct:attached_list 67 struct codec_list *next; member in struct:codec_list 141 while (a->next) videocodec_attach() 142 a = a->next; // find end videocodec_attach() 143 a->next = ptr; videocodec_attach() 155 h = h->next; videocodec_attach() 209 h->list = a->next; videocodec_detach() 213 prev->next = a->next; videocodec_detach() 224 a = a->next; videocodec_detach() 226 h = h->next; videocodec_detach() 258 while (h->next) videocodec_register() 259 h = h->next; // find the end videocodec_register() 260 h->next = ptr; videocodec_register() 301 codeclist_top = h->next; videocodec_unregister() 305 prev->next = h->next; videocodec_unregister() 313 h = h->next; videocodec_unregister() 344 a = a->next; proc_videocodecs_show() 346 h = h->next; proc_videocodecs_show()
|
/linux-4.4.14/arch/um/include/asm/ |
H A D | mmu_context.h | 50 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 55 if(prev != next){ switch_mm() 57 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm() 58 if(next != &init_mm) switch_mm() 59 __switch_mm(&next->context.id); switch_mm()
|
/linux-4.4.14/sound/isa/gus/ |
H A D | gus_mem.c | 55 nblock->next = pblock; snd_gf1_mem_xalloc() 60 nblock->prev->next = nblock; snd_gf1_mem_xalloc() 64 pblock = pblock->next; snd_gf1_mem_xalloc() 66 nblock->next = NULL; snd_gf1_mem_xalloc() 72 alloc->last->next = nblock; snd_gf1_mem_xalloc() 86 alloc->first = block->next; snd_gf1_mem_xfree() 87 if (block->next) snd_gf1_mem_xfree() 88 block->next->prev = NULL; snd_gf1_mem_xfree() 90 block->prev->next = block->next; snd_gf1_mem_xfree() 91 if (block->next) snd_gf1_mem_xfree() 92 block->next->prev = block->prev; snd_gf1_mem_xfree() 97 block->prev->next = NULL; snd_gf1_mem_xfree() 99 block->next->prev = block->prev; snd_gf1_mem_xfree() 101 block->prev->next = block->next; snd_gf1_mem_xfree() 113 for (block = alloc->first; block; block = block->next) { snd_gf1_mem_look() 129 for (block = alloc->first; block; block = block->next) snd_gf1_mem_share() 154 block->prev = block->next = NULL; snd_gf1_mem_find() 155 for (pblock = alloc->first, idx = 0; pblock; pblock = pblock->next) { snd_gf1_mem_find() 161 if (pblock->next) { snd_gf1_mem_find() 162 if (pblock->ptr + pblock->size == pblock->next->ptr) snd_gf1_mem_find() 164 if (pblock->next->ptr < boundary) snd_gf1_mem_find() 165 ptr2 = pblock->next->ptr; snd_gf1_mem_find() 280 nblock = block->next; snd_gf1_mem_done() 311 for (block = alloc->first, i = 0; block; block = block->next, i++) { snd_gf1_mem_info_read()
|
/linux-4.4.14/drivers/isdn/i4l/ |
H A D | isdn_net.h | 87 nd->queue = nd->queue->next; isdn_net_get_locked_lp() 94 nd->queue = nd->queue->next; isdn_net_get_locked_lp() 118 lp->last->next = nlp; isdn_net_add_to_bundle() 120 nlp->next = lp; isdn_net_add_to_bundle() 136 // printk(KERN_DEBUG "%s: lp:%s(%p) mlp:%s(%p) last(%p) next(%p) mndq(%p)\n", isdn_net_rm_from_bundle() 137 // __func__, lp->name, lp, master_lp->name, master_lp, lp->last, lp->next, master_lp->netdev->queue); isdn_net_rm_from_bundle() 139 lp->last->next = lp->next; isdn_net_rm_from_bundle() 140 lp->next->last = lp->last; isdn_net_rm_from_bundle() 142 master_lp->netdev->queue = lp->next; isdn_net_rm_from_bundle() 143 if (lp->next == lp) { /* last in queue */ isdn_net_rm_from_bundle() 147 lp->next = lp->last = lp; /* (re)set own pointers */ isdn_net_rm_from_bundle()
|
/linux-4.4.14/mm/kasan/ |
H A D | kasan_init.c | 66 unsigned long next; zero_pmd_populate() local 69 next = pmd_addr_end(addr, end); zero_pmd_populate() 80 zero_pte_populate(pmd, addr, next); zero_pmd_populate() 81 } while (pmd++, addr = next, addr != end); zero_pmd_populate() 88 unsigned long next; zero_pud_populate() local 91 next = pud_addr_end(addr, end); zero_pud_populate() 105 zero_pmd_populate(pud, addr, next); zero_pud_populate() 106 } while (pud++, addr = next, addr != end); zero_pud_populate() 121 unsigned long next; kasan_populate_zero_shadow() local 124 next = pgd_addr_end(addr, end); kasan_populate_zero_shadow() 150 zero_pud_populate(pgd, addr, next); kasan_populate_zero_shadow() 151 } while (pgd++, addr = next, addr != end); kasan_populate_zero_shadow()
|
/linux-4.4.14/drivers/isdn/hardware/eicon/ |
H A D | pr_pc.h | 27 word NextReq; /* pointer to next Req Buffer */ 28 word NextRc; /* pointer to next Rc Buffer */ 29 word NextInd; /* pointer to next Ind Buffer */ 46 word next; member in struct:__anon5602 56 word next; member in struct:__anon5603 65 word next; member in struct:__anon5604
|
/linux-4.4.14/drivers/acpi/ |
H A D | wakeup.c | 32 struct list_head *node, *next; acpi_enable_wakeup_devices() local 34 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_enable_wakeup_devices() 59 struct list_head *node, *next; acpi_disable_wakeup_devices() local 61 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_disable_wakeup_devices() 81 struct list_head *node, *next; acpi_wakeup_device_init() local 84 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_wakeup_device_init()
|
H A D | proc.c | 24 struct list_head *node, *next; acpi_system_wakeup_device_seq_show() local 29 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_system_wakeup_device_seq_show() 57 dev->physical_node_list.next) acpi_system_wakeup_device_seq_show() 98 struct list_head *node, *next; acpi_system_write_wakeup_device() local 111 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_system_write_wakeup_device()
|
/linux-4.4.14/fs/f2fs/ |
H A D | shrinker.c | 46 p = f2fs_list.next; f2fs_shrink_count() 52 p = p->next; f2fs_shrink_count() 67 p = p->next; f2fs_shrink_count() 87 p = f2fs_list.next; f2fs_shrink_scan() 96 p = p->next; f2fs_shrink_scan() 115 p = p->next; f2fs_shrink_scan()
|
/linux-4.4.14/drivers/acpi/acpica/ |
H A D | pstree.c | 107 arg = arg->common.next; acpi_ps_get_arg() 166 while (prev_arg->common.next) { acpi_ps_append_arg() 167 prev_arg = prev_arg->common.next; acpi_ps_append_arg() 169 prev_arg->common.next = arg; acpi_ps_append_arg() 180 arg = arg->common.next; acpi_ps_append_arg() 195 * DESCRIPTION: Get next op in tree (walking the tree in depth-first order) 203 union acpi_parse_object *next = NULL; acpi_ps_get_depth_next() local 215 next = acpi_ps_get_arg(op, 0); acpi_ps_get_depth_next() 216 if (next) { acpi_ps_get_depth_next() 217 return (next); acpi_ps_get_depth_next() 222 next = op->common.next; acpi_ps_get_depth_next() 223 if (next) { acpi_ps_get_depth_next() 224 return (next); acpi_ps_get_depth_next() 234 arg = arg->common.next; acpi_ps_get_depth_next() 244 if (parent->common.next) { acpi_ps_get_depth_next() 248 return (parent->common.next); acpi_ps_get_depth_next() 255 return (next); acpi_ps_get_depth_next()
|
H A D | pswalk.c | 65 union acpi_parse_object *next = NULL; acpi_ps_delete_parse_tree() local 80 next = acpi_ps_get_arg(op, 0); acpi_ps_delete_parse_tree() 81 if (next) { acpi_ps_delete_parse_tree() 85 op = next; acpi_ps_delete_parse_tree() 92 next = op->common.next; acpi_ps_delete_parse_tree() 102 if (next) { acpi_ps_delete_parse_tree() 103 op = next; acpi_ps_delete_parse_tree()
|
H A D | evgpeutil.c | 98 gpe_block = gpe_block->next; acpi_ev_walk_gpe_list() 101 gpe_xrupt_info = gpe_xrupt_info->next; acpi_ev_walk_gpe_list() 184 next_gpe_xrupt = next_gpe_xrupt->next; acpi_ev_get_gpe_xrupt_block() 201 while (next_gpe_xrupt->next) { acpi_ev_get_gpe_xrupt_block() 202 next_gpe_xrupt = next_gpe_xrupt->next; acpi_ev_get_gpe_xrupt_block() 205 next_gpe_xrupt->next = gpe_xrupt; acpi_ev_get_gpe_xrupt_block() 271 gpe_xrupt->previous->next = gpe_xrupt->next; acpi_ev_delete_gpe_xrupt() 275 acpi_gbl_gpe_xrupt_list_head = gpe_xrupt->next; acpi_ev_delete_gpe_xrupt() 278 if (gpe_xrupt->next) { acpi_ev_delete_gpe_xrupt() 279 gpe_xrupt->next->previous = gpe_xrupt->previous; acpi_ev_delete_gpe_xrupt() 310 struct acpi_gpe_notify_info *next; acpi_ev_delete_gpe_handlers() local 345 next = notify->next; acpi_ev_delete_gpe_handlers() 347 notify = next; acpi_ev_delete_gpe_handlers()
|
H A D | exdump.c | 156 {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(region.next), "Next"} 244 {ACPI_EXD_HDLR_LIST, ACPI_EXD_OFFSET(address_space.next), "Next"}, 257 {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[0]), 259 {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[1]), "Next Device Notify"} 373 union acpi_operand_object *next; acpi_ex_dump_object() local 471 next = start; acpi_ex_dump_object() 473 acpi_os_printf("%20s : %p", name, next); acpi_ex_dump_object() 474 if (next) { acpi_ex_dump_object() 477 (next), next->common.type); acpi_ex_dump_object() 479 while (next->common.next_object) { acpi_ex_dump_object() 480 if ((next->common.type == acpi_ex_dump_object() 482 data = next; acpi_ex_dump_object() 485 next = next->common.next_object; acpi_ex_dump_object() 486 acpi_os_printf("->%p(%s %2.2X)", next, acpi_ex_dump_object() 488 (next), acpi_ex_dump_object() 489 next->common.type); acpi_ex_dump_object() 491 if ((next == start) || (next == data)) { acpi_ex_dump_object() 505 next = start; acpi_ex_dump_object() 507 acpi_os_printf("%20s : %p", name, next); acpi_ex_dump_object() 508 if (next) { acpi_ex_dump_object() 511 (next), next->common.type); acpi_ex_dump_object() 513 while (next->address_space.next) { acpi_ex_dump_object() 514 if ((next->common.type == acpi_ex_dump_object() 516 data = next; acpi_ex_dump_object() 519 next = next->address_space.next; acpi_ex_dump_object() 520 acpi_os_printf("->%p(%s %2.2X)", next, acpi_ex_dump_object() 522 (next), acpi_ex_dump_object() 523 next->common.type); acpi_ex_dump_object() 525 if ((next == start) || (next == data)) { acpi_ex_dump_object() 539 next = start; acpi_ex_dump_object() 541 acpi_os_printf("%20s : %p", name, next); acpi_ex_dump_object() 542 if (next) { acpi_ex_dump_object() 545 (next), next->common.type); acpi_ex_dump_object() 547 while (next->region.next) { acpi_ex_dump_object() 548 if ((next->common.type == acpi_ex_dump_object() 550 data = next; acpi_ex_dump_object() 553 next = next->region.next; acpi_ex_dump_object() 554 acpi_os_printf("->%p(%s %2.2X)", next, acpi_ex_dump_object() 556 (next), acpi_ex_dump_object() 557 next->common.type); acpi_ex_dump_object() 559 if ((next == start) || (next == data)) { acpi_ex_dump_object()
|
H A D | utaddress.c | 106 range_info->next = acpi_gbl_address_range_list[space_id]; acpi_ut_add_address_range() 157 range_info->next; acpi_ut_remove_address_range() 159 prev->next = range_info->next; acpi_ut_remove_address_range() 176 range_info = range_info->next; acpi_ut_remove_address_range() 259 range_info = range_info->next; acpi_ut_check_address_range() 280 struct acpi_address_range *next; acpi_ut_delete_address_lists() local 287 next = acpi_gbl_address_range_list[i]; acpi_ut_delete_address_lists() 289 while (next) { acpi_ut_delete_address_lists() 290 range_info = next; acpi_ut_delete_address_lists() 291 next = range_info->next; acpi_ut_delete_address_lists()
|
/linux-4.4.14/arch/alpha/kernel/ |
H A D | err_common.c | 151 /* return the next header */ el_process_header_subpacket() 160 struct el_subpacket *next = NULL; el_process_subpacket_reg() 163 for (; h && h->class != header->class; h = h->next); el_process_subpacket_reg() 164 if (h) next = h->handler(header); el_process_subpacket_reg() 166 return next; 198 struct el_subpacket *next = NULL; el_process_subpacket() 205 next = el_process_header_subpacket(header); el_process_subpacket() 208 if (NULL == (next = el_process_subpacket_reg(header))) { el_process_subpacket() 217 return next; 226 for (a = subpacket_annotation_list; a; a = a->next) { el_annotate_subpacket() 287 for (; a->next != NULL; a = a->next) { cdl_register_subpacket_annotation() 295 a->next = new; cdl_register_subpacket_annotation() 297 new->next = NULL; cdl_register_subpacket_annotation() 309 for (; h->next != NULL; h = h->next) { cdl_register_subpacket_handler() 316 h->next = new; cdl_register_subpacket_handler() 318 new->next = NULL; cdl_register_subpacket_handler() 158 struct el_subpacket *next = NULL; el_process_subpacket_reg() local 196 struct el_subpacket *next = NULL; el_process_subpacket() local
|
H A D | gct.c | 39 if (node->next) gct6_find_nodes() 40 status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search); gct6_find_nodes()
|
/linux-4.4.14/kernel/ |
H A D | task_work.c | 5 static struct callback_head work_exited; /* all we need is ->next == NULL */ 35 work->next = head; task_work_add() 70 pprev = &work->next; task_work_cancel() 71 else if (cmpxchg(pprev, work, work->next) == work) task_work_cancel() 90 struct callback_head *work, *head, *next; task_work_run() local 114 next = work->next; task_work_run() 116 work = next; task_work_run()
|
/linux-4.4.14/arch/mips/mm/ |
H A D | gup.c | 104 unsigned long next; gup_pmd_range() local 111 next = pmd_addr_end(addr, end); gup_pmd_range() 126 if (!gup_huge_pmd(pmd, addr, next, write, pages,nr)) gup_pmd_range() 129 if (!gup_pte_range(pmd, addr, next, write, pages,nr)) gup_pmd_range() 132 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 170 unsigned long next; gup_pud_range() local 177 next = pud_addr_end(addr, end); gup_pud_range() 181 if (!gup_huge_pud(pud, addr, next, write, pages,nr)) gup_pud_range() 184 if (!gup_pmd_range(pud, addr, next, write, pages,nr)) gup_pud_range() 187 } while (pudp++, addr = next, addr != end); gup_pud_range() 201 unsigned long next; __get_user_pages_fast() local 236 next = pgd_addr_end(addr, end); __get_user_pages_fast() 239 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) __get_user_pages_fast() 241 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast() 268 unsigned long next; get_user_pages_fast() local 286 next = pgd_addr_end(addr, end); get_user_pages_fast() 289 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) get_user_pages_fast() 291 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
|
/linux-4.4.14/arch/nios2/include/asm/ |
H A D | switch_to.h | 17 #define switch_to(prev, next, last) \ 26 : "r" (prev), "r" (next) \
|
/linux-4.4.14/arch/x86/mm/ |
H A D | init_64.c | 73 unsigned long next; ident_pud_init() local 75 for (; addr < end; addr = next) { ident_pud_init() 79 next = (addr & PUD_MASK) + PUD_SIZE; ident_pud_init() 80 if (next > end) ident_pud_init() 81 next = end; ident_pud_init() 85 ident_pmd_init(info->pmd_flag, pmd, addr, next); ident_pud_init() 91 ident_pmd_init(info->pmd_flag, pmd, addr, next); ident_pud_init() 101 unsigned long next; kernel_ident_mapping_init() local 105 for (; addr < end; addr = next) { kernel_ident_mapping_init() 109 next = (addr & PGDIR_MASK) + PGDIR_SIZE; kernel_ident_mapping_init() 110 if (next > end) kernel_ident_mapping_init() 111 next = end; kernel_ident_mapping_init() 115 result = ident_pud_init(info, pud, addr, next); kernel_ident_mapping_init() 124 result = ident_pud_init(info, pud, addr, next); kernel_ident_mapping_init() 405 unsigned long pages = 0, next; phys_pte_init() local 411 for (i = pte_index(addr); i < PTRS_PER_PTE; i++, addr = next, pte++) { phys_pte_init() 412 next = (addr & PAGE_MASK) + PAGE_SIZE; phys_pte_init() 415 !e820_any_mapped(addr & PAGE_MASK, next, E820_RAM) && phys_pte_init() 416 !e820_any_mapped(addr & PAGE_MASK, next, E820_RESERVED_KERN)) phys_pte_init() 450 unsigned long pages = 0, next; phys_pmd_init() local 455 for (; i < PTRS_PER_PMD; i++, address = next) { phys_pmd_init() 460 next = (address & PMD_MASK) + PMD_SIZE; phys_pmd_init() 463 !e820_any_mapped(address & PMD_MASK, next, E820_RAM) && phys_pmd_init() 464 !e820_any_mapped(address & PMD_MASK, next, E820_RESERVED_KERN)) phys_pmd_init() 493 last_map_addr = next; phys_pmd_init() 506 last_map_addr = next; phys_pmd_init() 525 unsigned long pages = 0, next; phys_pud_init() local 529 for (; i < PTRS_PER_PUD; i++, addr = next) { phys_pud_init() 534 next = (addr & PUD_MASK) + PUD_SIZE; phys_pud_init() 537 !e820_any_mapped(addr & PUD_MASK, next, E820_RAM) && phys_pud_init() 538 !e820_any_mapped(addr & PUD_MASK, next, E820_RESERVED_KERN)) phys_pud_init() 566 last_map_addr = next; phys_pud_init() 579 last_map_addr = next; phys_pud_init() 604 unsigned long next, last_map_addr = end; kernel_physical_mapping_init() local 611 for (; start < end; start = next) { kernel_physical_mapping_init() 615 next = (start & PGDIR_MASK) + PGDIR_SIZE; kernel_physical_mapping_init() 722 magic = (unsigned long)page->lru.next; free_pagetable() 794 unsigned long next, pages = 0; remove_pte_table() local 800 for (; addr < end; addr = next, pte++) { remove_pte_table() 801 next = (addr + PAGE_SIZE) & PAGE_MASK; remove_pte_table() 802 if (next > end) remove_pte_table() 803 next = end; remove_pte_table() 818 IS_ALIGNED(next, PAGE_SIZE)) { remove_pte_table() 843 memset((void *)addr, PAGE_INUSE, next - addr); remove_pte_table() 866 unsigned long next, pages = 0; remove_pmd_table() local 872 for (; addr < end; addr = next, pmd++) { remove_pmd_table() 873 next = pmd_addr_end(addr, end); remove_pmd_table() 880 IS_ALIGNED(next, PMD_SIZE)) { remove_pmd_table() 891 memset((void *)addr, PAGE_INUSE, next - addr); remove_pmd_table() 909 remove_pte_table(pte_base, addr, next, direct); remove_pmd_table() 922 unsigned long next, pages = 0; remove_pud_table() local 928 for (; addr < end; addr = next, pud++) { remove_pud_table() 929 next = pud_addr_end(addr, end); remove_pud_table() 936 IS_ALIGNED(next, PUD_SIZE)) { remove_pud_table() 947 memset((void *)addr, PAGE_INUSE, next - addr); remove_pud_table() 965 remove_pmd_table(pmd_base, addr, next, direct); remove_pud_table() 977 unsigned long next; remove_pagetable() local 983 for (addr = start; addr < end; addr = next) { remove_pagetable() 984 next = pgd_addr_end(addr, end); remove_pagetable() 991 remove_pud_table(pud, addr, next, direct); remove_pagetable() 1242 unsigned long next; vmemmap_populate_hugepages() local 1247 for (addr = start; addr < end; addr = next) { vmemmap_populate_hugepages() 1248 next = pmd_addr_end(addr, end); vmemmap_populate_hugepages() 1285 vmemmap_verify((pte_t *)pmd, node, addr, next); vmemmap_populate_hugepages() 1289 if (vmemmap_populate_basepages(addr, next, node)) vmemmap_populate_hugepages() 1314 unsigned long next; register_page_bootmem_memmap() local 1321 for (; addr < end; addr = next) { register_page_bootmem_memmap() 1326 next = (addr + PAGE_SIZE) & PAGE_MASK; register_page_bootmem_memmap() 1333 next = (addr + PAGE_SIZE) & PAGE_MASK; register_page_bootmem_memmap() 1339 next = (addr + PAGE_SIZE) & PAGE_MASK; register_page_bootmem_memmap() 1352 next = pmd_addr_end(addr, end); register_page_bootmem_memmap()
|
/linux-4.4.14/net/ipv4/ |
H A D | ip_fragment.c | 314 struct sk_buff *xp = fp->next; ip_frag_reinit() 336 struct sk_buff *prev, *next; ip_frag_queue() local 405 next = NULL; ip_frag_queue() 409 for (next = qp->q.fragments; next != NULL; next = next->next) { ip_frag_queue() 410 if (FRAG_CB(next)->offset >= offset) ip_frag_queue() 412 prev = next; ip_frag_queue() 438 while (next && FRAG_CB(next)->offset < end) { ip_frag_queue() 439 int i = end - FRAG_CB(next)->offset; /* overlap is 'i' bytes */ ip_frag_queue() 441 if (i < next->len) { ip_frag_queue() 442 /* Eat head of the next overlapped fragment ip_frag_queue() 443 * and leave the loop. The next ones cannot overlap. ip_frag_queue() 445 if (!pskb_pull(next, i)) ip_frag_queue() 447 FRAG_CB(next)->offset += i; ip_frag_queue() 449 if (next->ip_summed != CHECKSUM_UNNECESSARY) ip_frag_queue() 450 next->ip_summed = CHECKSUM_NONE; ip_frag_queue() 453 struct sk_buff *free_it = next; ip_frag_queue() 458 next = next->next; ip_frag_queue() 461 prev->next = next; ip_frag_queue() 463 qp->q.fragments = next; ip_frag_queue() 474 skb->next = next; ip_frag_queue() 475 if (!next) ip_frag_queue() 478 prev->next = skb; ip_frag_queue() 544 head = prev->next; ip_frag_reasm() 549 fp->next = head->next; ip_frag_reasm() 550 if (!fp->next) ip_frag_reasm() 552 prev->next = fp; ip_frag_reasm() 555 head->next = qp->q.fragments->next; ip_frag_reasm() 586 clone->next = head->next; ip_frag_reasm() 587 head->next = clone; ip_frag_reasm() 600 skb_shinfo(head)->frag_list = head->next; ip_frag_reasm() 603 for (fp=head->next; fp; fp = fp->next) { ip_frag_reasm() 614 head->next = NULL; ip_frag_reasm()
|
/linux-4.4.14/arch/sparc/mm/ |
H A D | gup.c | 121 unsigned long next; gup_pmd_range() local 128 next = pmd_addr_end(addr, end); gup_pmd_range() 132 if (!gup_huge_pmd(pmdp, pmd, addr, next, gup_pmd_range() 135 } else if (!gup_pte_range(pmd, addr, next, write, gup_pmd_range() 138 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 146 unsigned long next; gup_pud_range() local 153 next = pud_addr_end(addr, end); gup_pud_range() 156 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) gup_pud_range() 158 } while (pudp++, addr = next, addr != end); gup_pud_range() 168 unsigned long next, flags; __get_user_pages_fast() local 182 next = pgd_addr_end(addr, end); __get_user_pages_fast() 185 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) __get_user_pages_fast() 187 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast() 198 unsigned long next; get_user_pages_fast() local 230 next = pgd_addr_end(addr, end); get_user_pages_fast() 233 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) get_user_pages_fast() 235 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
|
/linux-4.4.14/net/core/ |
H A D | dst.c | 70 struct dst_entry *dst, *next, head; dst_gc_task() local 74 next = dst_busy_list; dst_gc_task() 77 while ((dst = next) != NULL) { dst_gc_task() 78 next = dst->next; dst_gc_task() 79 prefetch(&next->next); dst_gc_task() 82 last->next = dst; dst_gc_task() 103 dst->next = next; dst_gc_task() 104 next = dst; dst_gc_task() 109 next = dst_garbage.list; dst_gc_task() 110 if (next) { dst_gc_task() 115 last->next = NULL; dst_gc_task() 116 dst_busy_list = head.next; dst_gc_task() 135 * if the next desired timer is more than 4 seconds in the dst_gc_task() 194 dst->next = NULL; dst_init() 236 dst->next = dst_garbage.list; __dst_free() 452 for (dst = dst_busy_list; dst; dst = dst->next) { dst_dev_event() 463 last->next = dst; dst_dev_event() 466 for (; dst; dst = dst->next) dst_dev_event()
|
/linux-4.4.14/drivers/scsi/pcmcia/ |
H A D | nsp_message.c | 22 * the next "msg in" if exists (no scsi phase changes). nsp_message_in() 45 /* catch a next signal */ nsp_message_in() 63 * the next "msg out" if exists (no scsi phase changes). nsp_message_out() 72 /* catch a next signal */ nsp_message_out()
|
/linux-4.4.14/drivers/md/ |
H A D | raid0.h | 5 sector_t zone_end; /* Start of the next zone (in sectors) */
|
/linux-4.4.14/drivers/net/appletalk/ |
H A D | ipddp.h | 20 struct ipddp_route *next; member in struct:ipddp_route
|
/linux-4.4.14/drivers/net/wireless/b43/ |
H A D | debugfs.h | 58 /* The cached address for the next mmio16read file read */ 60 /* The cached address for the next mmio32read file read */ 63 /* The cached address for the next shm16read file read */ 66 /* The cached address for the next shm32read file read */
|
/linux-4.4.14/arch/c6x/kernel/ |
H A D | switch_to.S | 17 * struct thread_info *next, 23 || MV .L1X B4,A5 ; next 45 ;; Switch to next SP 61 B .S2 B3 ; return in next E1
|
/linux-4.4.14/include/net/ |
H A D | atmclip.h | 33 struct clip_vcc *next; /* next VCC */ member in struct:clip_vcc 49 struct net_device *next; /* next CLIP interface */ member in struct:clip_priv
|
/linux-4.4.14/include/asm-generic/bitops/ |
H A D | find.h | 6 * find_next_bit - find the next set bit in a memory region 11 * Returns the bit number for the next set bit 20 * find_next_zero_bit - find the next cleared bit in a memory region 25 * Returns the bit number of the next zero bit
|
/linux-4.4.14/arch/sh/include/cpu-sh5/cpu/ |
H A D | mmu_context.h | 16 /* This has to be a common function because the next location to fill
|
/linux-4.4.14/arch/m68k/include/asm/ |
H A D | switch_to.h | 20 * Beware that resume now expects *next to be in d1 and the offset of 27 * pass prev in a0, next in a1 30 #define switch_to(prev,next,last) do { \ 32 register void *_next __asm__ ("a1") = (next); \
|
/linux-4.4.14/arch/h8300/include/asm/ |
H A D | switch_to.h | 20 * Beware that resume now expects *next to be in d1 and the offset of 27 * pass prev in a0, next in a1, offset of tss in d1, and whether 34 #define switch_to(prev, next, last) \ 45 "r" (&(next->thread)), \
|
/linux-4.4.14/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 132 struct sk_buff *prev, *next; lowpan_frag_queue() local 169 next = NULL; lowpan_frag_queue() 173 for (next = fq->q.fragments; next != NULL; next = next->next) { lowpan_frag_queue() 174 if (lowpan_802154_cb(next)->d_offset >= lowpan_frag_queue() 177 prev = next; lowpan_frag_queue() 182 skb->next = next; lowpan_frag_queue() 183 if (!next) lowpan_frag_queue() 186 prev->next = skb; lowpan_frag_queue() 236 head = prev->next; lowpan_frag_reasm() 242 fp->next = head->next; lowpan_frag_reasm() 243 if (!fp->next) lowpan_frag_reasm() 245 prev->next = fp; lowpan_frag_reasm() 248 head->next = fq->q.fragments->next; lowpan_frag_reasm() 269 clone->next = head->next; lowpan_frag_reasm() 270 head->next = clone; lowpan_frag_reasm() 285 for (fp = head->next; fp;) { lowpan_frag_reasm() 288 struct sk_buff *next = fp->next; lowpan_frag_reasm() local 300 fp = next; lowpan_frag_reasm() 304 head->next = NULL; lowpan_frag_reasm()
|
/linux-4.4.14/net/rose/ |
H A D | rose_route.c | 70 rose_node = rose_node->next; rose_add_node() 84 rose_neigh = rose_neigh->next; rose_add_node() 129 rose_neigh->next = rose_neigh_list; rose_add_node() 147 rose_tmpn = rose_tmpn->next; rose_add_node() 169 rose_node->next = NULL; rose_add_node() 171 rose_tmpp->next = rose_node; rose_add_node() 172 rose_node->next = NULL; rose_add_node() 176 rose_node->next = rose_node_list; rose_add_node() 179 rose_tmpp->next = rose_node; rose_add_node() 180 rose_node->next = rose_tmpn; rose_add_node() 210 rose_node_list = rose_node->next; rose_remove_node() 215 while (s != NULL && s->next != NULL) { rose_remove_node() 216 if (s->next == rose_node) { rose_remove_node() 217 s->next = rose_node->next; rose_remove_node() 222 s = s->next; rose_remove_node() 239 rose_neigh_list = rose_neigh->next; rose_remove_neigh() 247 while (s != NULL && s->next != NULL) { rose_remove_neigh() 248 if (s->next == rose_neigh) { rose_remove_neigh() 249 s->next = rose_neigh->next; rose_remove_neigh() 257 s = s->next; rose_remove_neigh() 275 rose_route_list = rose_route->next; rose_remove_route() 280 while (s != NULL && s->next != NULL) { rose_remove_route() 281 if (s->next == rose_route) { rose_remove_route() 282 s->next = rose_route->next; rose_remove_route() 287 s = s->next; rose_remove_route() 311 rose_node = rose_node->next; rose_del_node() 325 rose_neigh = rose_neigh->next; rose_del_node() 397 sn->next = rose_neigh_list; rose_add_loopback_neigh() 418 rose_node = rose_node->next; rose_add_loopback_node() 436 rose_node->next = rose_node_list; rose_add_loopback_node() 462 rose_node = rose_node->next; rose_del_loopback_node() 490 rose_neigh = rose_neigh->next; rose_rt_device_down() 499 rose_node = rose_node->next; rose_rt_device_down() 539 rose_route = rose_route->next; 566 rose_node = rose_node->next; rose_clear_routes() 573 rose_neigh = rose_neigh->next; rose_clear_routes() 664 for (rose_route = rose_route_list; rose_route != NULL; rose_route = rose_route->next) rose_route_free_lci() 684 for (node = rose_node_list; node != NULL; node = node->next) { rose_get_neigh() 695 for (node = rose_node_list; node != NULL; node = node->next) { rose_get_neigh() 783 s = rose_route->next; rose_del_route_by_neigh() 801 rose_route = rose_route->next; rose_del_route_by_neigh() 820 rose_neigh = rose_neigh->next; rose_link_failed() 841 for (rose_neigh = rose_neigh_list; rose_neigh != NULL; rose_neigh = rose_neigh->next) { rose_link_device_down() 886 rose_neigh = rose_neigh->next; rose_route_frame() 955 * Route it to the next in line if we have an entry for it. rose_route_frame() 1001 rose_route = rose_route->next; rose_route_frame() 1035 rose_route = rose_route->next; rose_route_frame() 1066 rose_route->next = rose_route_list; rose_route_frame() 1096 rose_node = rose_node->next, ++i); __acquires() 1106 : ((struct rose_node *)v)->next; rose_node_next() 1146 .next = rose_node_next, 1175 rose_neigh = rose_neigh->next, ++i); __acquires() 1185 : ((struct rose_neigh *)v)->next; rose_neigh_next() 1230 .next = rose_neigh_next, 1260 rose_route = rose_route->next, ++i); __acquires() 1270 : ((struct rose_route *)v)->next; rose_route_next() 1316 .next = rose_route_next, 1347 rose_neigh = rose_neigh->next; rose_rt_free() 1354 rose_node = rose_node->next; rose_rt_free() 1361 rose_route = rose_route->next; rose_rt_free()
|
/linux-4.4.14/drivers/staging/rtl8192u/ieee80211/ |
H A D | ieee80211_crypt.c | 45 for (ptr = ieee->crypt_deinit_list.next, n = ptr->next; ieee80211_crypt_deinit_entries() 46 ptr != &ieee->crypt_deinit_list; ptr = n, n = ptr->next) { ieee80211_crypt_deinit_entries() 136 for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) { ieee80211_unregister_crypto_ops() 167 for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) { ieee80211_get_crypto_ops() 229 for (ptr = hcrypt->algs.next, n = ptr->next; ptr != &hcrypt->algs; ieee80211_crypto_deinit() 230 ptr = n, n = ptr->next) { ieee80211_crypto_deinit()
|
/linux-4.4.14/drivers/staging/rdma/amso1100/ |
H A D | c2_alloc.c | 54 new_head->next = NULL; c2_alloc_mqsp_chunk() 57 /* build list where each index is the next free slot */ c2_alloc_mqsp_chunk() 79 struct sp_chunk *next; c2_free_mqsp_pool() local 82 next = root->next; c2_free_mqsp_pool() 85 root = next; c2_free_mqsp_pool() 99 } else if (head->next == NULL) { c2_alloc_mqsp() 100 if (c2_alloc_mqsp_chunk(c2dev, gfp_mask, &head->next) == c2_alloc_mqsp() 102 head = head->next; c2_alloc_mqsp() 109 head = head->next; c2_alloc_mqsp()
|
/linux-4.4.14/fs/nilfs2/ |
H A D | segbuf.h | 42 * @next: Block number of the next full segment 54 sector_t next; member in struct:nilfs_segsum_info 63 * @sb_nextnum: Index number of the next full segment 100 #define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next) 103 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next) 104 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head)) 111 (list_entry((head)->next, struct buffer_head, b_assoc_buffers)) 113 (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \ 115 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head)
|
/linux-4.4.14/arch/alpha/lib/ |
H A D | memset.S | 34 bis $17,$1,$17 /* E0 (p-c latency, next cycle) */ 35 sll $17,16,$1 /* E1 (p-c latency, next cycle) */ 37 bis $17,$1,$17 /* E0 (p-c latency, next cycle) */ 38 sll $17,32,$1 /* E1 (p-c latency, next cycle) */ 39 bis $17,$1,$17 /* E0 (p-c latency, next cycle) */
|
/linux-4.4.14/drivers/isdn/divert/ |
H A D | isdn_divert.c | 29 struct call_struc *next; /* pointer to next entry */ member in struct:call_struc 38 struct deflect_struc *next, *prev; member in struct:deflect_struc 48 static ulong next_id = 1; /* next info id */ 94 cs->prev->next = cs->next; /* forward link */ deflect_timer_expire() 96 divert_head = cs->next; deflect_timer_expire() 97 if (cs->next) deflect_timer_expire() 98 cs->next->prev = cs->prev; /* back link */ deflect_timer_expire() 193 cs->next = divert_head; cf_command() 216 cs = cs->next; deflect_extern_action() 289 ds1 = ds1->next; insertrule() 296 ds->next = NULL; /* end of chain */ insertrule() 298 ds->prev->next = ds; /* last forward */ insertrule() 303 ds->next = ds1; /* next entry */ insertrule() 330 ds = ds->next; deleterule() 341 ds = ds->next; deleterule() 349 if (ds->next) deleterule() 350 ds->next->prev = ds->prev; /* backward chain */ deleterule() 355 ds->prev->next = ds->next; /* forward chain */ deleterule() 357 table_head = ds->next; /* start of chain */ deleterule() 377 ds = ds->next; getruleptr() 395 for (dv = table_head; dv; dv = dv->next) { isdn_divert_icall() 520 cs->next = divert_head; isdn_divert_icall() 543 cs = cs->next; deleteprocs() 744 cs = cs->next; prot_stat_callback() 757 cs1->prev->next = cs1->next; /* forward link */ prot_stat_callback() 759 divert_head = cs1->next; prot_stat_callback() 760 if (cs1->next) prot_stat_callback() 761 cs1->next->prev = cs1->prev; /* back link */ prot_stat_callback() 809 cs = cs->next; isdn_divert_stat_callback() 813 cs1->prev->next = cs1->next; /* forward link */ isdn_divert_stat_callback() 815 divert_head = cs1->next; isdn_divert_stat_callback() 816 if (cs1->next) isdn_divert_stat_callback() 817 cs1->next->prev = cs1->prev; /* back link */ isdn_divert_stat_callback()
|
/linux-4.4.14/fs/omfs/ |
H A D | file.c | 37 u64 next, last; omfs_shrink_inode() local 45 next = inode->i_ino; omfs_shrink_inode() 52 bh = omfs_bread(inode->i_sb, next); omfs_shrink_inode() 61 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) omfs_shrink_inode() 69 last = next; omfs_shrink_inode() 70 next = be64_to_cpu(oe->e_next); omfs_shrink_inode() 89 if (next == ~0) omfs_shrink_inode() 92 bh = omfs_bread(inode->i_sb, next); omfs_shrink_inode() 141 /* trivially grow current extent, if next block is not taken */ omfs_grow_extent() 221 sector_t next, offset; omfs_get_block() local 239 next = inode->i_ino; omfs_get_block() 243 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) omfs_get_block() 247 next = be64_to_cpu(oe->e_next); omfs_get_block() 262 if (next == ~0) omfs_get_block() 266 bh = omfs_bread(inode->i_sb, next); omfs_get_block()
|
/linux-4.4.14/drivers/char/agp/ |
H A D | frontend.c | 59 curr = curr->next; agp_find_mem_by_key() 69 struct agp_memory *next; agp_remove_from_pool() local 75 next = temp->next; agp_remove_from_pool() 79 prev->next = next; agp_remove_from_pool() 80 if (next != NULL) agp_remove_from_pool() 81 next->prev = prev; agp_remove_from_pool() 85 if (next != NULL) agp_remove_from_pool() 86 next->prev = NULL; agp_remove_from_pool() 88 agp_fe.current_controller->pool = next; agp_remove_from_pool() 207 temp->next = prev; agp_insert_into_pool() 224 curr = curr->next; agp_find_private() 238 priv->next = prev; agp_insert_file_private() 244 struct agp_file_private *next; agp_remove_file_private() local 247 next = priv->next; agp_remove_file_private() 251 prev->next = next; agp_remove_file_private() 253 if (next != NULL) agp_remove_file_private() 254 next->prev = prev; agp_remove_file_private() 257 if (next != NULL) agp_remove_file_private() 258 next->prev = NULL; agp_remove_file_private() 260 agp_fe.file_priv_list = next; agp_remove_file_private() 302 controller = controller->next; agp_find_controller_by_pid() 325 controller->next = prev_controller; agp_insert_controller() 353 client = client->next; agp_remove_all_clients() 367 memory = memory->next; agp_remove_all_memory() 378 next_controller = controller->next; agp_remove_controller() 381 prev_controller->next = next_controller; agp_remove_controller() 419 clients = clients->next; agp_controller_make_current() 441 clients = clients->next; agp_controller_release_current() 467 client = client->next; agp_find_client_in_controller() 482 controller = controller->next; agp_find_controller_for_client() 504 client->next = prev_client; agp_insert_client() 542 next_client = client->next; agp_remove_client() 545 prev_client->next = next_client; agp_remove_client()
|
/linux-4.4.14/net/atm/ |
H A D | mpoa_caches.c | 47 entry = entry->next; in_cache_get() 68 entry = entry->next; in_cache_get_with_mask() 89 entry = entry->next; in_cache_get_by_vcc() 111 entry->next = client->in_cache; in_cache_add_entry() 201 entry->prev->next = entry->next; in_cache_remove_entry() 203 client->in_cache = entry->next; in_cache_remove_entry() 204 if (entry->next != NULL) in_cache_remove_entry() 205 entry->next->prev = entry->prev; in_cache_remove_entry() 237 next_entry = entry->next; clear_count_and_expired() 266 entry = entry->next; /* Entry in hold down */ check_resolving_entries() 279 entry = entry->next; check_resolving_entries() 294 entry = entry->next; check_resolving_entries() 320 entry = entry->next; refresh_entries() 346 entry = entry->next; eg_cache_get_by_cache_id() 367 entry = entry->next; eg_cache_get_by_tag() 389 entry = entry->next; eg_cache_get_by_vcc() 409 entry = entry->next; eg_cache_get_by_src_ip() 436 entry->prev->next = entry->next; eg_cache_remove_entry() 438 client->eg_cache = entry->next; eg_cache_remove_entry() 439 if (entry->next != NULL) eg_cache_remove_entry() 440 entry->next->prev = entry->prev; eg_cache_remove_entry() 474 entry->next = client->eg_cache; eg_cache_add_entry() 513 next_entry = entry->next; clear_expired()
|
/linux-4.4.14/scripts/dtc/ |
H A D | livetree.c | 41 new->next = *labels; add_label() 79 assert(first->next == NULL); chain_property() 81 first->next = list; chain_property() 89 struct property *next; reverse_properties() local 92 next = p->next; reverse_properties() 93 p->next = head; reverse_properties() 95 p = next; reverse_properties() 154 new_node->proplist = new_prop->next; merge_nodes() 155 new_prop->next = NULL; merge_nodes() 231 prop->next = NULL; add_property() 235 p = &((*p)->next); add_property() 249 prop = prop->next; delete_property_by_name() 314 assert(first->next == NULL); chain_reserve_entry() 316 first->next = list; chain_reserve_entry() 325 new->next = NULL; add_reserve_entry() 330 for (last = list; last->next; last = last->next) add_reserve_entry() 333 last->next = new; add_reserve_entry() 602 ri = ri->next) sort_reserve_entries() 612 ri = ri->next) sort_reserve_entries() 619 tbl[i]->next = tbl[i+1]; sort_reserve_entries() 620 tbl[n-1]->next = NULL; sort_reserve_entries() 655 tbl[i]->next = tbl[i+1]; sort_properties() 656 tbl[n-1]->next = NULL; sort_properties()
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | mmu_context.h | 32 static inline void contextidr_thread_switch(struct task_struct *next) contextidr_thread_switch() argument 38 : "r" (task_pid_nr(next))); contextidr_thread_switch() 41 static inline void contextidr_thread_switch(struct task_struct *next) contextidr_thread_switch() argument 127 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 132 if (prev == next) switch_mm() 139 if (next == &init_mm) { switch_mm() 144 check_and_switch_context(next, cpu); switch_mm() 148 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
|
H A D | spinlock_types.h | 29 u16 next; member in struct:__anon306 33 u16 next;
|
/linux-4.4.14/drivers/base/ |
H A D | map.c | 21 struct probe *next; member in struct:kobj_map::probe 60 s = &(*s)->next; kobj_map() 61 p->next = *s; kobj_map() 81 for (s = &domain->probes[index % 255]; *s; s = &(*s)->next) { kobj_unmap() 84 *s = p->next; kobj_unmap() 103 for (p = domain->probes[MAJOR(dev) % 255]; p; p = p->next) { kobj_lookup()
|
/linux-4.4.14/fs/jffs2/ |
H A D | nodelist.c | 43 new->next = (*prev)->next; jffs2_add_fd_to_list() 53 prev = &((*prev)->next); jffs2_add_fd_to_list() 55 new->next = *prev; jffs2_add_fd_to_list() 73 struct jffs2_node_frag *next = frag_next(frag); jffs2_truncate_fragtree() local 77 frag = next; jffs2_truncate_fragtree() 384 or next node REF_NORMAL, as appropriate. */ jffs2_add_full_dnode_to_inode() 395 struct jffs2_node_frag *next = frag_next(newfrag); jffs2_add_full_dnode_to_inode() local 397 if (next) { jffs2_add_full_dnode_to_inode() 399 if (next->node) jffs2_add_full_dnode_to_inode() 400 mark_ref_normal(next->node->raw); jffs2_add_full_dnode_to_inode() 427 ret = ret->next; jffs2_get_ino_cache() 449 prev = &(*prev)->next; jffs2_add_ino_cache() 451 new->next = *prev; jffs2_add_ino_cache() 470 prev = &(*prev)->next; jffs2_del_ino_cache() 473 *prev = old->next; jffs2_del_ino_cache() 490 struct jffs2_inode_cache *this, *next; jffs2_free_ino_caches() local 495 next = this->next; jffs2_free_ino_caches() 498 this = next; jffs2_free_ino_caches() 507 struct jffs2_raw_node_ref *this, *next; jffs2_free_raw_node_refs() local 513 next = this[REFS_PER_BLOCK].next_in_ino; jffs2_free_raw_node_refs() 515 next = NULL; jffs2_free_raw_node_refs() 518 this = next; jffs2_free_raw_node_refs() 528 struct rb_node *next; jffs2_lookup_node_frag() local 534 next = fragtree->rb_node; jffs2_lookup_node_frag() 536 while(next) { jffs2_lookup_node_frag() 537 frag = rb_entry(next, struct jffs2_node_frag, rb); jffs2_lookup_node_frag() 543 next = frag->rb.rb_right; jffs2_lookup_node_frag() 545 next = frag->rb.rb_left; jffs2_lookup_node_frag() 567 struct jffs2_node_frag *frag, *next; jffs2_kill_fragtree() local 570 rbtree_postorder_for_each_entry_safe(frag, next, root, rb) { rbtree_postorder_for_each_entry_safe() 734 pr_crit("next %p (0x%08x-0x%08x)\n", __jffs2_ref_totlen() 738 pr_crit("No next ref. jeb->last_node is %p\n", __jffs2_ref_totlen()
|
/linux-4.4.14/net/sched/ |
H A D | sch_sfq.c | 99 sfq_index next; member in struct:sfq_head 107 sfq_index next; /* next slot in sfq RR chain */ member in struct:sfq_slot 209 n = q->dep[qlen].next; sfq_link() 211 slot->dep.next = n; sfq_link() 214 q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */ sfq_link() 220 n = q->slots[x].dep.next; \ 222 sfq_dep_head(q, p)->next = n; \ 261 skb->prev->next = (struct sk_buff *)slot; slot_dequeue_tail() 262 skb->next = skb->prev = NULL; slot_dequeue_tail() 271 slot->skblist_next = skb->next; slot_dequeue_head() 272 skb->next->prev = (struct sk_buff *)slot; slot_dequeue_head() 273 skb->next = skb->prev = NULL; slot_dequeue_head() 287 skb->next = (struct sk_buff *)slot; slot_queue_add() 288 slot->skblist_prev->next = skb; slot_queue_add() 302 x = q->dep[d].next; sfq_drop() 318 x = q->tail->next; sfq_drop() 320 q->tail->next = slot->next; sfq_drop() 368 x = q->dep[0].next; /* get a free slot */ sfq_enqueue() 447 slot->next = x; sfq_enqueue() 449 slot->next = q->tail->next; sfq_enqueue() 450 q->tail->next = x; sfq_enqueue() 489 a = q->tail->next; sfq_dequeue() 505 next_a = slot->next; sfq_dequeue() 510 q->tail->next = next_a; sfq_dequeue() 565 x = q->dep[0].next; /* get a free slot */ sfq_rehash() 589 slot->next = x; sfq_rehash() 591 slot->next = q->tail->next; sfq_rehash() 592 q->tail->next = x; sfq_rehash() 721 q->dep[i].next = i + SFQ_MAX_FLOWS; sfq_init()
|
/linux-4.4.14/fs/ |
H A D | pnode.c | 16 /* return the next shared peer mount of @p */ next_peer() 19 return list_entry(p->mnt_share.next, struct mount, mnt_share); next_peer() 24 return list_entry(p->mnt_slave_list.next, struct mount, mnt_slave); first_slave() 29 return list_entry(p->mnt_slave.next, struct mount, mnt_slave); next_slave() 137 * get the next mount in the propagation tree. 157 struct mount *next = next_peer(m); propagation_next() local 158 return (next == origin) ? NULL : next; propagation_next() 159 } else if (m->mnt_slave.next != &master->mnt_slave_list) propagation_next() 171 struct mount *next; next_group() local 174 next = next_peer(m); next_group() 176 if (next == origin) next_group() 178 } else if (m->mnt_slave.next != &next->mnt_slave) next_group() 180 m = next; next_group() 185 if (m->mnt_slave.next != &master->mnt_slave_list) next_group() 190 if (master->mnt_slave.next == &m->mnt_slave) next_group()
|
H A D | char_dev.c | 32 struct char_device_struct *next; member in struct:char_device_struct 54 for (cd = chrdevs[offset]; cd; cd = cd->next) chrdev_show() 108 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) __register_chrdev_region() 135 cd->next = *cp; __register_chrdev_region() 152 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) __unregister_chrdev_region() 159 *cp = cd->next; __unregister_chrdev_region() 178 dev_t n, next; register_chrdev_region() local 180 for (n = from; n < to; n = next) { register_chrdev_region() 181 next = MKDEV(MAJOR(n)+1, 0); register_chrdev_region() 182 if (next > to) register_chrdev_region() 183 next = to; register_chrdev_region() 185 next - n, name); register_chrdev_region() 192 for (n = from; n < to; n = next) { register_chrdev_region() 193 next = MKDEV(MAJOR(n)+1, 0); register_chrdev_region() 194 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); register_chrdev_region() 288 dev_t n, next; unregister_chrdev_region() local 290 for (n = from; n < to; n = next) { unregister_chrdev_region() 291 next = MKDEV(MAJOR(n)+1, 0); unregister_chrdev_region() 292 if (next > to) unregister_chrdev_region() 293 next = to; unregister_chrdev_region() 294 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); unregister_chrdev_region() 413 inode = container_of(cdev->list.next, struct inode, i_devices); cdev_purge()
|
/linux-4.4.14/arch/s390/mm/ |
H A D | gup.c | 108 unsigned long next; gup_pmd_range() local 118 next = pmd_addr_end(addr, end); gup_pmd_range() 138 if (!gup_huge_pmd(pmdp, pmd, addr, next, gup_pmd_range() 141 } else if (!gup_pte_range(pmdp, pmd, addr, next, gup_pmd_range() 144 } while (pmdp++, addr = next, addr != end); gup_pmd_range() 152 unsigned long next; gup_pud_range() local 162 next = pud_addr_end(addr, end); gup_pud_range() 165 if (!gup_pmd_range(pudp, pud, addr, next, write, pages, nr)) gup_pud_range() 167 } while (pudp++, addr = next, addr != end); gup_pud_range() 181 unsigned long next, flags; __get_user_pages_fast() local 203 next = pgd_addr_end(addr, end); __get_user_pages_fast() 206 if (!gup_pud_range(pgdp, pgd, addr, next, write, pages, &nr)) __get_user_pages_fast() 208 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
|
/linux-4.4.14/drivers/isdn/capi/ |
H A D | capilib.c | 12 struct capilib_msgidqueue *next; member in struct:capilib_msgidqueue 40 np->msgidpool[i].next = np->msgidfree; mq_init() 50 np->msgidfree = mq->next; mq_enqueue() 52 mq->next = NULL; mq_enqueue() 54 np->msgidlast->next = mq; mq_enqueue() 65 for (pp = &np->msgidqueue; *pp; pp = &(*pp)->next) { mq_dequeue() 68 *pp = mq->next; mq_dequeue() 71 mq->next = np->msgidfree; mq_dequeue()
|
/linux-4.4.14/drivers/isdn/hysdn/ |
H A D | hysdn_init.c | 95 card->next = NULL; /*end of chain */ hysdn_pci_init_one() 97 card_last->next = card; /* pointer to next card */ hysdn_pci_init_one() 129 card_root = card_root->next; hysdn_pci_remove_one() 135 if (tmp->next == card) hysdn_pci_remove_one() 136 tmp->next = card->next; hysdn_pci_remove_one() 138 tmp = tmp->next; hysdn_pci_remove_one()
|
/linux-4.4.14/net/6lowpan/ |
H A D | nhc.h | 19 * @_idlen: len for the next header id and mask, should be always the same. 53 * struct lowpan_nhc - hold 6lowpan next hdr compression ifnformation 56 * @name: name of the specific next header compression 57 * @nexthdr: next header value of the protocol which should be compressed. 61 * @len: the length of the next header id and mask. 62 * @setup: callback to setup fill the next header id value and mask. 124 * lowpan_nhc_add - register a next header compression to framework 131 * lowpan_nhc_del - delete a next header compression from framework
|
/linux-4.4.14/kernel/debug/kdb/ |
H A D | Makefile | 17 /^\#/{next} \ 18 /^[ \t]*$$/{next} \
|
/linux-4.4.14/include/asm-generic/ |
H A D | switch_to.h | 25 #define switch_to(prev, next, last) \ 27 ((last) = __switch_to((prev), (next))); \
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | unwind_i.h | 49 struct unw_table *next; /* must be first member! */ member in struct:unw_table 71 UNW_WHERE_GR_SAVE /* register is saved in next general register */ 83 struct unw_reg_state *next; /* next (outer) element on state stack */ member in struct:unw_reg_state 88 struct unw_labeled_state *next; /* next labeled state (or NULL) */ member in struct:unw_labeled_state 110 u8 gr_save_loc; /* next general register to use for saving a register */ 161 unsigned short hint; /* hint for next script to try (or -1) */
|
/linux-4.4.14/arch/avr32/kernel/ |
H A D | switch_to.S | 15 /* Switch thread context from "prev" to "next", returning "last" 18 * r10 : &next->thread
|
/linux-4.4.14/tools/include/asm-generic/bitops/ |
H A D | find.h | 6 * find_next_bit - find the next set bit in a memory region 11 * Returns the bit number for the next set bit
|
/linux-4.4.14/drivers/gpu/drm/atmel-hlcdc/ |
H A D | atmel_hlcdc_layer.c | 135 dscr->next); atmel_hlcdc_layer_update_apply() 151 dscr->next); atmel_hlcdc_layer_update_apply() 368 upd->next = upd->pending ? 0 : 1; atmel_hlcdc_layer_update_start() 370 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_start() 416 upd->slots[upd->next].configs, atmel_hlcdc_layer_update_start() 429 atmel_hlcdc_layer_update_reset(layer, upd->next); atmel_hlcdc_layer_update_rollback() 430 upd->next = -1; atmel_hlcdc_layer_update_rollback() 445 if (upd->next < 0 || upd->next > 1) atmel_hlcdc_layer_update_set_fb() 454 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_set_fb() 483 if (upd->next < 0 || upd->next > 1) atmel_hlcdc_layer_update_cfg() 489 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_cfg() 502 if (upd->next < 0 || upd->next > 1) atmel_hlcdc_layer_update_commit() 505 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_commit() 515 upd->pending = upd->next; atmel_hlcdc_layer_update_commit() 516 upd->next = -1; atmel_hlcdc_layer_update_commit() 524 upd->next = -1; atmel_hlcdc_layer_update_commit() 544 dscr->next = dma_addr + (i * sizeof(*dscr)); atmel_hlcdc_layer_dma_init() 564 dma->dscrs[0].next); atmel_hlcdc_layer_dma_cleanup() 595 upd->next = -1; atmel_hlcdc_layer_update_init()
|
/linux-4.4.14/block/partitions/ |
H A D | msdos.c | 114 * table start). The second is a pointer to the next logical partition 139 if (state->next == state->limit) parse_extended() 152 * the 2nd entry is the next extended partition, or empty, parse_extended() 163 sector_t offs, size, next; parse_extended() local 172 next = this_sector + offs; parse_extended() 176 if (next < first_sector) parse_extended() 178 if (next + size > first_sector + first_size) parse_extended() 182 put_partition(state, state->next, next, size); parse_extended() 183 set_info(state, state->next, disksig); parse_extended() 185 state->parts[state->next].flags = ADDPART_FLAG_RAID; parse_extended() 187 if (++state->next == state->limit) parse_extended() 195 * It should be a link to the next logical partition. parse_extended() 248 for (i = 0; i < max_nparts && state->next < state->limit; i++) { parse_solaris_x86() 258 put_partition(state, state->next++, parse_solaris_x86() 297 if (state->next == state->limit) parse_bsd() 310 put_partition(state, state->next++, bsd_start, bsd_size); parse_bsd() 376 if (state->next == state->limit) parse_unixware() 380 put_partition(state, state->next++, parse_unixware() 420 if (state->next == state->limit) parse_minix() 424 put_partition(state, state->next++, parse_minix() 527 state->next = 5; msdos_partition()
|
/linux-4.4.14/drivers/pcmcia/ |
H A D | sa1111_generic.h | 7 struct sa1111_pcmcia_socket *next; member in struct:sa1111_pcmcia_socket
|
/linux-4.4.14/arch/um/os-Linux/ |
H A D | irq.c | 47 irq_fd = irq_fd->next; os_waiting_for_events() 106 if (*last_irq_ptr2 == &old_fd->next) os_free_irq_by_cb() 109 *prev = (*prev)->next; os_free_irq_by_cb() 115 prev = &(*prev)->next; os_free_irq_by_cb()
|
/linux-4.4.14/arch/unicore32/kernel/ |
H A D | time.c | 39 unsigned long next, oscr; puv3_osmr0_set_next_event() local 42 next = readl(OST_OSCR) + delta; puv3_osmr0_set_next_event() 43 writel(next, OST_OSMR0); puv3_osmr0_set_next_event() 46 return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0; puv3_osmr0_set_next_event()
|
/linux-4.4.14/include/linux/raid/ |
H A D | xor.h | 10 struct xor_block_template *next; member in struct:xor_block_template
|
/linux-4.4.14/firmware/dsp56k/ |
H A D | bootstrap.asm | 68 next jclr #0,X:<<M_HSR,* label 87 _load_P jmp next 91 _load_X jmp next 95 _load_Y jmp next
|
/linux-4.4.14/tools/perf/util/ |
H A D | rblist.c | 107 struct rb_node *pos, *next = rb_first(&rblist->entries); rblist__delete() local 109 while (next) { rblist__delete() 110 pos = next; rblist__delete() 111 next = rb_next(pos); rblist__delete()
|
/linux-4.4.14/drivers/media/pci/cobalt/ |
H A D | cobalt-omnitek.c | 167 dma_addr_t next = desc->bus; descriptor_list_create() local 176 WARN_ON(next & 3); descriptor_list_create() 221 next += sizeof(struct sg_dma_descriptor); descriptor_list_create() 222 d->next_h = (u32)((u64)next >> 32); descriptor_list_create() 223 d->next_l = (u32)next | descriptor_list_create() 265 next += sizeof(struct sg_dma_descriptor); descriptor_list_create() 275 d->next_h = (u32)((u64)next >> 32); descriptor_list_create() 276 d->next_l = (u32)next | (to_pci ? WRITE_TO_PCI : 0); descriptor_list_create() 284 struct sg_dma_desc_info *next) descriptor_list_chain() 289 if (next == NULL) { descriptor_list_chain() 293 d->next_h = (u32)((u64)next->bus >> 32); descriptor_list_chain() 294 d->next_l = (u32)next->bus | direction | INTERRUPT_ENABLE; descriptor_list_chain() 283 descriptor_list_chain(struct sg_dma_desc_info *this, struct sg_dma_desc_info *next) descriptor_list_chain() argument
|
/linux-4.4.14/fs/jfs/ |
H A D | jfs_txnmgr.h | 48 lid_t next; /* index of first tlock of tid */ member in struct:tblock 96 lid_t next; /* 2: index next lockword on tid locklist member in struct:tlock 97 * next lockword on freelist 171 lid_t next; /* 2: next linelock */ member in struct:linelock 187 lid_t next; /* 2: */ member in struct:xtlock 212 * N.B. next field may be set by linelock, and should not 214 * N.B. index of the first pxdlock specifies index of next 218 lid_t next; /* 2: */ member in struct:maplock 221 u8 index; /* 2: next free maplock index */ 246 lid_t next; /* 2: */ member in struct:xdlistlock
|