Searched refs:next (Results 1 - 200 of 6358) sorted by relevance

1234567891011>>

/linux-4.4.14/lib/
H A Dlist_debug.c19 * the prev/next entries already!
24 struct list_head *next) __list_add()
26 WARN(next->prev != prev, __list_add()
27 "list_add corruption. next->prev should be " __list_add()
28 "prev (%p), but was %p. (next=%p).\n", __list_add()
29 prev, next->prev, next); __list_add()
30 WARN(prev->next != next, __list_add()
31 "list_add corruption. prev->next should be " __list_add()
32 "next (%p), but was %p. (prev=%p).\n", __list_add()
33 next, prev->next, prev); __list_add()
34 WARN(new == prev || new == next, __list_add()
35 "list_add double add: new=%p, prev=%p, next=%p.\n", __list_add()
36 new, prev, next); __list_add()
37 next->prev = new; __list_add()
38 new->next = next; __list_add()
40 prev->next = new; __list_add()
46 struct list_head *prev, *next; __list_del_entry() local
49 next = entry->next; __list_del_entry()
51 if (WARN(next == LIST_POISON1, __list_del_entry()
52 "list_del corruption, %p->next is LIST_POISON1 (%p)\n", __list_del_entry()
57 WARN(prev->next != entry, __list_del_entry()
58 "list_del corruption. prev->next should be %p, " __list_del_entry()
59 "but was %p\n", entry, prev->next) || __list_del_entry()
60 WARN(next->prev != entry, __list_del_entry()
61 "list_del corruption. next->prev should be %p, " __list_del_entry()
62 "but was %p\n", entry, next->prev)) __list_del_entry()
65 __list_del(prev, next); __list_del_entry()
78 entry->next = LIST_POISON1; list_del()
87 struct list_head *prev, struct list_head *next) __list_add_rcu()
89 WARN(next->prev != prev, __list_add_rcu()
90 "list_add_rcu corruption. next->prev should be prev (%p), but was %p. (next=%p).\n", __list_add_rcu()
91 prev, next->prev, next); __list_add_rcu()
92 WARN(prev->next != next, __list_add_rcu()
93 "list_add_rcu corruption. prev->next should be next (%p), but was %p. (prev=%p).\n", __list_add_rcu()
94 next, prev->next, prev); __list_add_rcu()
95 new->next = next; __list_add_rcu()
98 next->prev = new; __list_add_rcu()
22 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
86 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
H A Dplist.c36 WARN(n->prev != p || p->next != n, plist_check_prev_next()
39 "next: %p, n: %p, p: %p\n", plist_check_prev_next()
40 t, t->next, t->prev, plist_check_prev_next()
41 p, p->next, p->prev, plist_check_prev_next()
42 n, n->next, n->prev); plist_check_prev_next()
47 struct list_head *prev = top, *next = top->next; plist_check_list() local
49 plist_check_prev_next(top, prev, next); plist_check_list()
50 while (next != top) { plist_check_list()
51 prev = next; plist_check_list()
52 next = prev->next; plist_check_list()
53 plist_check_prev_next(top, prev, next); plist_check_list()
95 iter = list_entry(iter->prio_list.next, plist_add()
118 if (node->node_list.next != &head->node_list) { plist_del()
119 struct plist_node *next; plist_del() local
121 next = list_entry(node->node_list.next, plist_del()
124 /* add the next plist_node into prio_list */ plist_del()
125 if (list_empty(&next->prio_list)) plist_del()
126 list_add(&next->prio_list, &node->prio_list); plist_del()
204 BUG_ON(prio_pos->prio_list.next != &node_pos->prio_list); plist_test_check()
209 BUG_ON(prio_pos->prio_list.next != &first->prio_list); plist_test_check()
H A Dioremap.c75 unsigned long next; ioremap_pmd_range() local
82 next = pmd_addr_end(addr, end); ioremap_pmd_range()
85 ((next - addr) == PMD_SIZE) && ioremap_pmd_range()
91 if (ioremap_pte_range(pmd, addr, next, phys_addr + addr, prot)) ioremap_pmd_range()
93 } while (pmd++, addr = next, addr != end); ioremap_pmd_range()
101 unsigned long next; ioremap_pud_range() local
108 next = pud_addr_end(addr, end); ioremap_pud_range()
111 ((next - addr) == PUD_SIZE) && ioremap_pud_range()
117 if (ioremap_pmd_range(pud, addr, next, phys_addr + addr, prot)) ioremap_pud_range()
119 } while (pud++, addr = next, addr != end); ioremap_pud_range()
128 unsigned long next; ioremap_page_range() local
137 next = pgd_addr_end(addr, end); ioremap_page_range()
138 err = ioremap_pud_range(pgd, addr, next, phys_addr+addr, prot); ioremap_page_range()
141 } while (pgd++, addr = next, addr != end); ioremap_page_range()
H A Dlist_sort.c29 tail->next = a; merge()
30 a = a->next; merge()
32 tail->next = b; merge()
33 b = b->next; merge()
35 tail = tail->next; merge()
37 tail->next = a?:b; merge()
38 return head.next; merge()
60 tail->next = a; merge_and_restore_back_links()
62 a = a->next; merge_and_restore_back_links()
64 tail->next = b; merge_and_restore_back_links()
66 b = b->next; merge_and_restore_back_links()
68 tail = tail->next; merge_and_restore_back_links()
70 tail->next = a ? : b; merge_and_restore_back_links()
80 (*cmp)(priv, tail->next, tail->next); merge_and_restore_back_links()
82 tail->next->prev = tail; merge_and_restore_back_links()
83 tail = tail->next; merge_and_restore_back_links()
84 } while (tail->next); merge_and_restore_back_links()
86 tail->next = head; merge_and_restore_back_links()
119 head->prev->next = NULL; list_sort()
120 list = head->next; list_sort()
124 list = list->next; list_sort()
125 cur->next = NULL; list_sort()
245 for (cur = head.next; cur->next != &head; cur = cur->next) { list_sort_test()
249 if (cur->next->prev != cur) { list_sort_test()
254 cmp_result = cmp(NULL, cur, cur->next); list_sort_test()
261 el1 = container_of(cur->next, struct debug_el, list); list_sort_test()
H A Dtimerqueue.c59 if (!head->next || node->expires.tv64 < head->next->expires.tv64) { timerqueue_add()
60 head->next = node; timerqueue_add()
79 /* update next pointer */ timerqueue_del()
80 if (head->next == node) { timerqueue_del()
83 head->next = rbn ? timerqueue_del()
88 return head->next != NULL; timerqueue_del()
103 struct rb_node *next; timerqueue_iterate_next() local
107 next = rb_next(&node->node); timerqueue_iterate_next()
108 if (!next) timerqueue_iterate_next()
110 return container_of(next, struct timerqueue_node, node); timerqueue_iterate_next()
H A Dllist.c44 new_last->next = first = ACCESS_ONCE(head->first); llist_add_batch()
61 * llist_add) sequence in another user may change @head->first->next,
67 struct llist_node *entry, *old_entry, *next; llist_del_first() local
74 next = READ_ONCE(entry->next); llist_del_first()
75 entry = cmpxchg(&head->first, old_entry, next); llist_del_first()
97 head = head->next; llist_reverse_order()
98 tmp->next = new_head; llist_reverse_order()
H A Dklist.c23 * next element in the list. It returns NULL if there are no more items.
26 * the next klist_node. It then drops the lock and returns.
368 * klist_next - Ante up next node in list.
372 * node, if there was one. Grab the next node, increment its reference
373 * count, drop the lock, and return that next node.
379 struct klist_node *next; klist_next() local
384 next = to_klist_node(last->n_node.next); klist_next()
388 next = to_klist_node(i->i_klist->k_list.next); klist_next()
391 while (next != to_klist_node(&i->i_klist->k_list)) { klist_next()
392 if (likely(!knode_dead(next))) { klist_next()
393 kref_get(&next->n_ref); klist_next()
394 i->i_cur = next; klist_next()
397 next = to_klist_node(next->n_node.next); klist_next()
/linux-4.4.14/arch/score/include/asm/
H A Dswitch_to.h4 extern void *resume(void *last, void *next, void *next_ti);
6 #define switch_to(prev, next, last) \
8 (last) = resume(prev, next, task_thread_info(next)); \
H A Dmmu_context.h72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
78 if ((next->context ^ asid_cache) & ASID_VERSION_MASK) switch_mm()
79 get_new_mmu_context(next); switch_mm()
81 pevn_set(next->context); switch_mm()
82 TLBMISS_HANDLER_SETUP_PGD(next->pgd); switch_mm()
102 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
107 get_new_mmu_context(next); activate_mm()
108 pevn_set(next->context); activate_mm()
109 TLBMISS_HANDLER_SETUP_PGD(next->pgd); activate_mm()
/linux-4.4.14/tools/firewire/
H A Dlist.h2 struct list *next, *prev; member in struct:list
8 list->next = list; list_init()
15 return list->next == list; list_empty()
22 new_link->next = link; list_insert()
23 new_link->prev->next = new_link; list_insert()
24 new_link->next->prev = new_link; list_insert()
36 list_insert(list->next, new_link); list_prepend()
42 link->prev->next = link->next; list_remove()
43 link->next->prev = link->prev; list_remove()
50 list_entry((list)->next, type, member)
56 list_entry((elm)->member.next, typeof(*elm), member)
/linux-4.4.14/arch/parisc/include/asm/
H A Dswitch_to.h8 #define switch_to(prev, next, last) do { \
9 (last) = _switch_to(prev, next); \
H A Dmmu_context.h52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
55 if (prev != next) { switch_mm()
56 mtctl(__pa(next->pgd), 25); switch_mm()
57 load_context(next->context); switch_mm()
63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
73 BUG_ON(next == &init_mm); /* Should never happen */ activate_mm()
75 if (next->context == 0) activate_mm()
76 next->context = alloc_sid(); activate_mm()
78 switch_mm(prev,next,current); activate_mm()
/linux-4.4.14/kernel/locking/
H A Dosq_lock.c32 * Get a stable @node->next pointer, either for unlock() or unqueue() purposes.
40 struct optimistic_spin_node *next = NULL; osq_wait_next() local
63 * We must xchg() the @node->next value, because if we were to osq_wait_next()
65 * @node->next might complete Step-A and think its @prev is osq_wait_next()
70 * wait for a new @node->next from its Step-C. osq_wait_next()
72 if (node->next) { osq_wait_next()
73 next = xchg(&node->next, NULL); osq_wait_next()
74 if (next) osq_wait_next()
81 return next; osq_wait_next()
87 struct optimistic_spin_node *prev, *next; osq_lock() local
92 node->next = NULL; osq_lock()
107 WRITE_ONCE(prev->next, node); osq_lock()
133 * Undo our @prev->next assignment; this will make @prev's osq_lock()
134 * unlock()/unqueue() wait for a next pointer since @lock points to us osq_lock()
139 if (prev->next == node && osq_lock()
140 cmpxchg(&prev->next, node, NULL) == node) osq_lock()
161 * Step - B -- stabilize @next osq_lock()
163 * Similar to unlock(), wait for @node->next or move @lock from @node osq_lock()
167 next = osq_wait_next(lock, node, prev); osq_lock()
168 if (!next) osq_lock()
174 * @prev is stable because its still waiting for a new @prev->next osq_lock()
175 * pointer, @next is stable because our @node->next pointer is NULL and osq_lock()
179 WRITE_ONCE(next->prev, prev); osq_lock()
180 WRITE_ONCE(prev->next, next); osq_lock()
187 struct optimistic_spin_node *node, *next; osq_unlock() local
201 next = xchg(&node->next, NULL); osq_unlock()
202 if (next) { osq_unlock()
203 WRITE_ONCE(next->locked, 1); osq_unlock()
207 next = osq_wait_next(lock, node, NULL); osq_unlock()
208 if (next) osq_unlock()
209 WRITE_ONCE(next->locked, 1); osq_unlock()
H A Dmcs_spinlock.h18 struct mcs_spinlock *next; member in struct:mcs_spinlock
68 node->next = NULL; mcs_spin_lock()
88 WRITE_ONCE(prev->next, node); mcs_spin_lock()
101 struct mcs_spinlock *next = READ_ONCE(node->next); mcs_spin_unlock() local
103 if (likely(!next)) { mcs_spin_unlock()
109 /* Wait until the next pointer is set */ mcs_spin_unlock()
110 while (!(next = READ_ONCE(node->next))) mcs_spin_unlock()
114 /* Pass lock to next waiter. */ mcs_spin_unlock()
115 arch_mcs_spin_unlock_contended(&next->locked); mcs_spin_unlock()
/linux-4.4.14/arch/s390/include/asm/
H A Dswitch_to.h31 #define switch_to(prev,next,last) do { \
37 if (next->mm) { \
38 update_cr_regs(next); \
40 restore_access_regs(&next->thread.acrs[0]); \
41 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \
43 prev = __switch_to(prev,next); \
H A Dmmu_context.h85 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
90 S390_lowcore.user_asce = next->context.asce; switch_mm()
91 if (prev == next) switch_mm()
94 cpumask_set_cpu(cpu, &next->context.cpu_attach_mask); switch_mm()
98 atomic_inc(&next->context.attach_count); switch_mm()
128 struct mm_struct *next) activate_mm()
130 switch_mm(prev, next, current); activate_mm()
131 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); activate_mm()
132 set_user_asce(next); activate_mm()
127 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
/linux-4.4.14/arch/sh/include/asm/
H A Dswitch_to_64.h21 struct task_struct *next,
24 #define switch_to(prev,next,last) \
26 if (last_task_used_math != next) { \
27 struct pt_regs *regs = next->thread.uregs; \
30 last = sh64_switch_to(prev, &prev->thread, next, \
31 &next->thread); \
H A Dswitch_to_32.h64 struct task_struct *next);
69 #define switch_to(prev, next, last) \
81 if (is_dsp_enabled(next)) \
82 __restore_dsp(next); \
87 __ts5 = (u32 *)next; \
88 __ts6 = (u32 *)&next->thread.sp; \
89 __ts7 = next->thread.pc; \
H A Dmmu_context.h119 struct mm_struct *next, switch_mm()
124 if (likely(prev != next)) { switch_mm()
125 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm()
126 set_TTB(next->pgd); switch_mm()
127 activate_context(next, cpu); switch_mm()
129 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) switch_mm()
130 activate_context(next, cpu); switch_mm()
133 #define activate_mm(prev, next) switch_mm((prev),(next),NULL)
118 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
/linux-4.4.14/arch/blackfin/include/asm/
H A Dswitch_to.h21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next);
24 #define switch_to(prev,next,last) \
28 memcpy (L1_SCRATCH_TASK_INFO, &task_thread_info(next)->l1_task_info, \
30 (last) = resume (prev, next); \
33 #define switch_to(prev, next, last) \
35 (last) = resume(prev, next); \
H A Dmmu.h12 struct sram_list_struct *next; member in struct:sram_list_struct
/linux-4.4.14/arch/cris/include/asm/
H A Dswitch_to.h8 extern struct task_struct *resume(struct task_struct *prev, struct task_struct *next, int);
9 #define switch_to(prev,next,last) last = resume(prev,next, \
H A Dmmu_context.h9 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
14 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
19 switch_mm(prev, next, NULL); activate_mm()
/linux-4.4.14/tools/perf/ui/
H A Dprogress.c19 if (p->curr >= p->next) { ui_progress__update()
20 p->next += p->step; ui_progress__update()
28 p->next = p->step = total / 16; ui_progress__init()
H A Dprogress.h10 u64 curr, next, step, total; member in struct:ui_progress
/linux-4.4.14/arch/unicore32/include/asm/
H A Dswitch_to.h17 * switch_to(prev, next) should switch from task `prev' to `next'
18 * `prev' will never be the same as `next'. schedule() itself
24 #define switch_to(prev, next, last) \
27 task_thread_info(next)); \
H A Dmmu_context.h49 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
54 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) switch_mm()
55 cpu_switch_mm(next->pgd, next); switch_mm()
59 #define activate_mm(prev, next) switch_mm(prev, next, NULL)
/linux-4.4.14/sound/pci/ctxfi/
H A Dctimap.c31 entry->next = entry->addr; input_mapper_add()
53 pos = head->next;
60 entry->next = pos_ent->addr;
62 pre_ent->next = entry->addr;
71 struct list_head *next, *pre, *head; input_mapper_delete() local
80 next = (entry->list.next == head) ? head->next : entry->list.next; input_mapper_delete()
84 entry->next = entry->addr = entry->user = entry->slot = 0; input_mapper_delete()
91 next_ent = list_entry(next, struct imapper, list); input_mapper_delete()
93 pre_ent->next = next_ent->addr; input_mapper_delete()
106 pos = head->next; free_input_mapper_list()
H A Dctimap.h28 unsigned short next; /* the next input mapper ram id */ member in struct:imapper
/linux-4.4.14/ipc/
H A Dmsgutil.c27 * The next 2 defines are here bc this is the only file
43 struct msg_msgseg *next; member in struct:msg_msgseg
44 /* the next part of the message follows immediately */
62 msg->next = NULL; alloc_msg()
66 pseg = &msg->next; alloc_msg()
74 seg->next = NULL; alloc_msg()
75 pseg = &seg->next; alloc_msg()
101 for (seg = msg->next; seg != NULL; seg = seg->next) { load_msg()
132 for (dst_pseg = dst->next, src_pseg = src->next; copy_msg()
134 dst_pseg = dst_pseg->next, src_pseg = src_pseg->next) { copy_msg()
161 for (seg = msg->next; seg != NULL; seg = seg->next) { store_msg()
177 seg = msg->next; free_msg()
180 struct msg_msgseg *tmp = seg->next; free_msg()
/linux-4.4.14/tools/usb/usbip/libsrc/
H A Dlist.h13 * sometimes we already know the next/prev entries and we can
19 struct list_head *next, *prev; member in struct:list_head
29 list->next = list; INIT_LIST_HEAD()
37 * the prev/next entries already!
41 struct list_head *next) __list_add()
43 next->prev = new; __list_add()
44 new->next = next; __list_add()
46 prev->next = new; __list_add()
59 __list_add(new, head, head->next); list_add()
63 * Delete a list entry by making the prev/next entries
67 * the prev/next entries already!
69 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument
71 next->prev = prev; __list_del()
72 prev->next = next; __list_del()
87 __list_del(entry->prev, entry->next); __list_del_entry()
92 __list_del(entry->prev, entry->next); list_del()
93 entry->next = LIST_POISON1; list_del()
111 for (pos = (head)->next; pos != (head); pos = pos->next)
120 for (pos = (head)->next, n = pos->next; pos != (head); \
121 pos = n, n = pos->next)
39 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
H A Dnames.c44 struct vendor *next; member in struct:vendor
50 struct product *next; member in struct:product
56 struct class *next; member in struct:class
62 struct subclass *next; member in struct:subclass
68 struct protocol *next; member in struct:protocol
74 struct genericstrtable *next; member in struct:genericstrtable
106 for (; v; v = v->next) names_vendor()
117 for (; p; p = p->next) names_product()
128 for (; c; c = c->next) names_class()
139 for (; s; s = s->next) names_subclass()
152 for (; p; p = p->next) names_protocol()
161 struct pool *next; member in struct:pool
181 p->next = pool_head; my_malloc()
201 pool = pool->next; names_free()
212 for (; v; v = v->next) new_vendor()
220 v->next = vendors[h]; new_vendor()
232 for (; p; p = p->next) new_product()
241 p->next = products[h]; new_product()
252 for (; c; c = c->next) new_class()
260 c->next = classes[h]; new_class()
271 for (; s; s = s->next) new_subclass()
280 s->next = subclasses[h]; new_subclass()
293 for (; p; p = p->next) new_protocol()
304 p->next = protocols[h]; new_protocol()
/linux-4.4.14/drivers/scsi/sym53c8xx_2/
H A Dsym_malloc.c77 while (!h[j].next) { ___sym_malloc()
79 h[j].next = (m_link_p) M_GET_MEM_CLUSTER(); ___sym_malloc()
80 if (h[j].next) ___sym_malloc()
81 h[j].next->next = NULL; ___sym_malloc()
87 a = h[j].next; ___sym_malloc()
89 h[j].next = h[j].next->next; ___sym_malloc()
93 h[j].next = (m_link_p) (a+s); ___sym_malloc()
94 h[j].next->next = NULL; ___sym_malloc()
133 ((m_link_p) a)->next = h[i].next; ___sym_mfree()
134 h[i].next = (m_link_p) a; ___sym_mfree()
140 while (q->next && q->next != (m_link_p) b) { ___sym_mfree()
141 q = q->next; ___sym_mfree()
143 if (!q->next) { ___sym_mfree()
144 ((m_link_p) a)->next = h[i].next; ___sym_mfree()
145 h[i].next = (m_link_p) a; ___sym_mfree()
148 q->next = q->next->next; ___sym_mfree()
235 vbp->next = mp->vtob[hc]; ___get_dma_mem_cluster()
253 vbpp = &(*vbpp)->next; ___free_dma_mem_cluster()
256 *vbpp = (*vbpp)->next; ___free_dma_mem_cluster()
268 for (mp = mp0.next; ___get_dma_pool()
270 mp = mp->next); ___get_dma_pool()
284 mp->next = mp0.next; ___cre_dma_pool()
285 mp0.next = mp; ___cre_dma_pool()
295 m_pool_p *pp = &mp0.next; ___del_dma_pool()
298 pp = &(*pp)->next; ___del_dma_pool()
300 *pp = (*pp)->next; ___del_dma_pool()
371 vp = vp->next; __vtobus()
/linux-4.4.14/drivers/thunderbolt/
H A Dcap.c30 && cap->extended_short.next == 0 tb_cap_is_long()
45 int next; tb_cap_next() local
51 next = cap->basic.next; tb_cap_next()
59 next = cap->basic.next; tb_cap_next()
61 next = cap->extended_short.next; tb_cap_next()
63 next = cap->extended_long.next; tb_cap_next()
69 if (next == offset) tb_cap_next()
71 return next; tb_cap_next()
/linux-4.4.14/arch/xtensa/include/asm/
H A Dswitch_to.h15 extern void *_switch_to(void *last, void *next);
17 #define switch_to(prev,next,last) \
19 (last) = _switch_to(prev, next); \
H A Dnommu_context.h18 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
22 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
H A Dmmu_context.h73 * Start new asid cycle; continue counting with next get_new_mmu_context()
123 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
127 int migrated = next->context.cpu != cpu; switch_mm()
131 next->context.cpu = cpu; switch_mm()
133 if (migrated || prev != next) switch_mm()
134 activate_context(next, cpu); switch_mm()
137 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
/linux-4.4.14/arch/microblaze/include/asm/
H A Dswitch_to.h16 struct thread_info *next);
18 #define switch_to(prev, next, last) \
21 task_thread_info(next)); \
/linux-4.4.14/drivers/pnp/isapnp/
H A Dcompat.c33 list = from ? from->global_list.next : pnp_cards.next; pnp_find_card()
40 list = list->next; pnp_find_card()
56 list = pnp_global.next; pnp_find_dev()
58 list = from->global_list.next; pnp_find_dev()
66 list = list->next; pnp_find_dev()
71 list = card->devices.next; pnp_find_dev()
73 list = from->card_list.next; pnp_find_dev()
82 list = list->next; pnp_find_dev()
/linux-4.4.14/arch/arc/include/asm/
H A Dswitch_to.h31 #define switch_to(prev, next, last) \
33 ARC_FPU_PREV(prev, next); \
34 last = __switch_to(prev, next);\
35 ARC_FPU_NEXT(next); \
/linux-4.4.14/arch/avr32/include/asm/
H A Dswitch_to.h18 #define ocd_switch(prev, next) \
21 ocd_write(PID, next->pid); \
24 #define ocd_switch(prev, next)
28 * switch_to(prev, next, last) should switch from task `prev' to task
29 * `next'. `prev' will never be the same as `next'.
41 #define switch_to(prev, next, last) \
43 ocd_switch(prev, next); \
45 &next->thread.cpu_context); \
H A Dmmu_context.h109 struct mm_struct *next, switch_mm()
112 if (likely(prev != next)) { switch_mm()
113 unsigned long __pgdir = (unsigned long)next->pgd; switch_mm()
116 activate_context(next); switch_mm()
122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
108 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
/linux-4.4.14/tools/power/cpupower/lib/
H A Dcpufreq.c80 struct cpufreq_available_governors *tmp, *next; cpufreq_put_available_governors() local
87 next = tmp->next; cpufreq_put_available_governors()
91 tmp = next; cpufreq_put_available_governors()
104 struct cpufreq_available_frequencies *tmp, *next; cpufreq_put_available_frequencies() local
111 next = tmp->next; cpufreq_put_available_frequencies()
113 tmp = next; cpufreq_put_available_frequencies()
125 struct cpufreq_affected_cpus *tmp, *next; cpufreq_put_affected_cpus() local
132 next = tmp->next; cpufreq_put_affected_cpus()
134 tmp = next; cpufreq_put_affected_cpus()
192 struct cpufreq_stats *tmp, *next; cpufreq_put_stats() local
199 next = tmp->next; cpufreq_put_stats()
201 tmp = next; cpufreq_put_stats()
/linux-4.4.14/block/
H A Dcmdline-parser.c42 char *next = strchr(++partdef, ')'); parse_subpart() local
44 if (!next) { parse_subpart()
50 length = min_t(int, next - partdef, parse_subpart()
55 partdef = ++next; parse_subpart()
92 char *next; parse_parts() local
104 next = strchr(bdevdef, ':'); parse_parts()
105 if (!next) { parse_parts()
110 length = min_t(int, next - bdevdef, sizeof(newparts->name) - 1); parse_parts()
117 while (next && *(++next)) { parse_parts()
118 bdevdef = next; parse_parts()
119 next = strchr(bdevdef, ','); parse_parts()
121 length = (!next) ? (sizeof(buf) - 1) : parse_parts()
122 min_t(int, next - bdevdef, sizeof(buf) - 1); parse_parts()
168 char *next; cmdline_parts_parse() local
173 next = pbuf = buf = kstrdup(cmdline, GFP_KERNEL); cmdline_parts_parse()
179 while (next && *pbuf) { cmdline_parts_parse()
180 next = strchr(pbuf, ';'); cmdline_parts_parse()
181 if (next) cmdline_parts_parse()
182 *next = '\0'; cmdline_parts_parse()
188 if (next) cmdline_parts_parse()
189 pbuf = ++next; cmdline_parts_parse()
H A Dblk-merge.c41 * If the next starting sector would be misaligned, stop the discard at blk_bio_discard_split()
554 struct request *next) ll_merge_requests_fn()
558 req->biotail->bi_seg_back_size + next->bio->bi_seg_front_size; ll_merge_requests_fn()
564 if (req_no_special_merge(req) || req_no_special_merge(next)) ll_merge_requests_fn()
567 if (req_gap_back_merge(req, next->bio)) ll_merge_requests_fn()
573 if ((blk_rq_sectors(req) + blk_rq_sectors(next)) > ll_merge_requests_fn()
577 total_phys_segments = req->nr_phys_segments + next->nr_phys_segments; ll_merge_requests_fn()
578 if (blk_phys_contig_segment(q, req->biotail, next->bio)) { ll_merge_requests_fn()
581 if (next->nr_phys_segments == 1) ll_merge_requests_fn()
582 next->biotail->bi_seg_back_size = seg_size; ll_merge_requests_fn()
589 if (blk_integrity_merge_rq(q, req, next) == false) ll_merge_requests_fn()
648 struct request *next) attempt_merge()
650 if (!rq_mergeable(req) || !rq_mergeable(next)) attempt_merge()
653 if (!blk_check_merge_flags(req->cmd_flags, next->cmd_flags)) attempt_merge()
659 if (blk_rq_pos(req) + blk_rq_sectors(req) != blk_rq_pos(next)) attempt_merge()
662 if (rq_data_dir(req) != rq_data_dir(next) attempt_merge()
663 || req->rq_disk != next->rq_disk attempt_merge()
664 || req_no_special_merge(next)) attempt_merge()
668 !blk_write_same_mergeable(req->bio, next->bio)) attempt_merge()
673 * from next to rq and release next. merge_requests_fn attempt_merge()
677 if (!ll_merge_requests_fn(q, req, next)) attempt_merge()
686 if ((req->cmd_flags | next->cmd_flags) & REQ_MIXED_MERGE || attempt_merge()
688 (next->cmd_flags & REQ_FAILFAST_MASK)) { attempt_merge()
690 blk_rq_set_mixed_merge(next); attempt_merge()
699 if (time_after(req->start_time, next->start_time)) attempt_merge()
700 req->start_time = next->start_time; attempt_merge()
702 req->biotail->bi_next = next->bio; attempt_merge()
703 req->biotail = next->biotail; attempt_merge()
705 req->__data_len += blk_rq_bytes(next); attempt_merge()
707 elv_merge_requests(q, req, next); attempt_merge()
710 * 'next' is going away, so update stats accordingly attempt_merge()
712 blk_account_io_merge(next); attempt_merge()
714 req->ioprio = ioprio_best(req->ioprio, next->ioprio); attempt_merge()
715 if (blk_rq_cpu_valid(next)) attempt_merge()
716 req->cpu = next->cpu; attempt_merge()
718 /* owner-ship of bio passed from next to req */ attempt_merge()
719 next->bio = NULL; attempt_merge()
720 __blk_put_request(q, next); attempt_merge()
726 struct request *next = elv_latter_request(q, rq); attempt_back_merge() local
728 if (next) attempt_back_merge()
729 return attempt_merge(q, rq, next); attempt_back_merge()
745 struct request *next) blk_attempt_req_merge()
747 return attempt_merge(q, rq, next); blk_attempt_req_merge()
553 ll_merge_requests_fn(struct request_queue *q, struct request *req, struct request *next) ll_merge_requests_fn() argument
647 attempt_merge(struct request_queue *q, struct request *req, struct request *next) attempt_merge() argument
744 blk_attempt_req_merge(struct request_queue *q, struct request *rq, struct request *next) blk_attempt_req_merge() argument
H A Dnoop-iosched.c16 struct request *next) noop_merged_requests()
18 list_del_init(&next->queuelist); noop_merged_requests()
57 if (rq->queuelist.next == &nd->queue) noop_latter_request()
15 noop_merged_requests(struct request_queue *q, struct request *rq, struct request *next) noop_merged_requests() argument
/linux-4.4.14/net/ax25/
H A Dax25_iface.c38 struct listen_struct *next; member in struct:listen_struct
51 ap->next = protocol_list; ax25_register_pid()
68 protocol_list = protocol->next; ax25_protocol_release()
72 while (protocol != NULL && protocol->next != NULL) { ax25_protocol_release()
73 if (protocol->next->pid == pid) { ax25_protocol_release()
74 protocol->next = protocol->next->next; ax25_protocol_release()
78 protocol = protocol->next; ax25_protocol_release()
118 listen->next = listen_list; ax25_listen_register()
139 listen_list = listen->next; ax25_listen_release()
145 while (listen != NULL && listen->next != NULL) { ax25_listen_release()
146 if (ax25cmp(&listen->next->callsign, callsign) == 0 && listen->next->dev == dev) { ax25_listen_release()
147 s = listen->next; ax25_listen_release()
148 listen->next = listen->next->next; ax25_listen_release()
154 listen = listen->next; ax25_listen_release()
167 for (protocol = protocol_list; protocol != NULL; protocol = protocol->next) ax25_protocol_function()
182 for (listen = listen_list; listen != NULL; listen = listen->next) ax25_listen_mine()
209 for (protocol = protocol_list; protocol != NULL; protocol = protocol->next) ax25_protocol_is_registered()
H A Dax25_dev.c40 for (ax25_dev = ax25_dev_list; ax25_dev != NULL; ax25_dev = ax25_dev->next) ax25_addr_ax25dev()
87 ax25_dev->next = ax25_dev_list; ax25_dev_device_up()
112 for (s = ax25_dev_list; s != NULL; s = s->next) ax25_dev_device_down()
117 ax25_dev_list = s->next; ax25_dev_device_down()
124 while (s != NULL && s->next != NULL) { ax25_dev_device_down()
125 if (s->next == ax25_dev) { ax25_dev_device_down()
126 s->next = ax25_dev->next; ax25_dev_device_down()
133 s = s->next; ax25_dev_device_down()
193 ax25_dev = ax25_dev->next; ax25_dev_free()
H A Dax25_route.c53 ax25_rt = ax25_rt->next; ax25_rt_device_down()
57 ax25_route_list = s->next; ax25_rt_device_down()
61 for (t = ax25_route_list; t != NULL; t = t->next) { ax25_rt_device_down()
62 if (t->next == s) { ax25_rt_device_down()
63 t->next = s->next; ax25_rt_device_down()
109 ax25_rt = ax25_rt->next; ax25_rt_add()
135 ax25_rt->next = ax25_route_list; ax25_rt_add()
161 ax25_rt = ax25_rt->next; ax25_rt_del()
165 ax25_route_list = s->next; ax25_rt_del()
168 for (t = ax25_route_list; t != NULL; t = t->next) { ax25_rt_del()
169 if (t->next == s) { ax25_rt_del()
170 t->next = s->next; ax25_rt_del()
216 ax25_rt = ax25_rt->next; ax25_rt_opt()
262 for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) { __acquires()
275 ((struct ax25_route *) v)->next; ax25_rt_seq_next()
328 .next = ax25_rt_seq_next,
364 for (ax25_rt = ax25_route_list; ax25_rt != NULL; ax25_rt = ax25_rt->next) { ax25_get_route()
499 ax25_rt = ax25_rt->next; ax25_rt_free()
/linux-4.4.14/arch/ia64/include/asm/
H A Dswitch_to.h44 #define __switch_to(prev,next,last) do { \
47 if (IA64_HAS_EXTRA_STATE(next)) \
48 ia64_load_extra(next); \
49 ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \
50 (last) = ia64_switch_to((next)); \
60 # define switch_to(prev,next,last) do { \
66 __switch_to(prev, next, last); \
67 /* "next" in old context is "current" in new context */ \
76 # define switch_to(prev,next,last) __switch_to(prev, next, last)
H A Dmmu_context.h36 unsigned int next; /* next context number to use */ member in struct:ia64_ctx
39 /* call wrap_mmu_context when next >= max */
91 if (ia64_ctx.next >= ia64_ctx.limit) { get_mmu_context()
92 ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, get_mmu_context()
93 ia64_ctx.max_ctx, ia64_ctx.next); get_mmu_context()
95 ia64_ctx.max_ctx, ia64_ctx.next); get_mmu_context()
96 if (ia64_ctx.next >= ia64_ctx.max_ctx) get_mmu_context()
99 mm->context = context = ia64_ctx.next++; get_mmu_context()
185 activate_mm (struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
191 ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd)); activate_mm()
192 activate_context(next); activate_mm()
/linux-4.4.14/scripts/kconfig/
H A Dlist.h24 struct list_head *next, *prev; member in struct:list_head
49 for (pos = list_entry((head)->next, typeof(*pos), member); \
51 pos = list_entry(pos->member.next, typeof(*pos), member))
61 for (pos = list_entry((head)->next, typeof(*pos), member), \
62 n = list_entry(pos->member.next, typeof(*pos), member); \
64 pos = n, n = list_entry(n->member.next, typeof(*n), member))
72 return head->next == head; list_empty()
79 * the prev/next entries already!
83 struct list_head *next) __list_add()
85 next->prev = _new; __list_add()
86 _new->next = next; __list_add()
88 prev->next = _new; __list_add()
105 * Delete a list entry by making the prev/next entries
109 * the prev/next entries already!
111 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument
113 next->prev = prev; __list_del()
114 prev->next = next; __list_del()
127 __list_del(entry->prev, entry->next); list_del()
128 entry->next = (struct list_head*)LIST_POISON1; list_del()
81 __list_add(struct list_head *_new, struct list_head *prev, struct list_head *next) __list_add() argument
H A Dkxgettext.c43 goto next; escape()
50 next: escape()
64 struct file_line *next; member in struct:file_line
78 self->next = NULL; file_line__new()
86 struct message *next; member in struct:message
109 self->next = NULL; message__new()
127 m = m->next; mesage__find()
142 fl->next = self->files; message__add_file_line()
163 m->next = message__list; message__add()
184 for (child = menu->list; child != NULL; child = child->next) menu_build_message_list()
198 fl = fl->next; message__print_file_lineno()
202 fl = fl->next; message__print_file_lineno()
224 m = m->next; menu__xgettext()
/linux-4.4.14/arch/um/include/shared/
H A Daio.h18 struct aio_context *next; member in struct:aio_context
22 .next = NULL }
H A Dirq_user.h12 struct irq_fd *next; member in struct:irq_fd
/linux-4.4.14/arch/arm/include/asm/
H A Dspinlock_types.h15 u16 next; member in struct:__anon191::__anon192::__raw_tickets
19 u16 next;
H A Dswitch_to.h19 * switch_to(prev, next) should switch from task `prev' to `next'
20 * `prev' will never be the same as `next'. schedule() itself
25 #define switch_to(prev,next,last) \
28 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
H A Dmmu_context.h93 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
116 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
128 !cpumask_empty(mm_cpumask(next)) && switch_mm()
129 !cpumask_test_cpu(cpu, mm_cpumask(next))) switch_mm()
132 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { switch_mm()
133 check_and_switch_context(next, tsk); switch_mm()
/linux-4.4.14/arch/frv/include/asm/
H A Dswitch_to.h18 * switch_to(prev, next) should switch from task `prev' to `next'
19 * `prev' will never be the same as `next'.
27 #define switch_to(prev, next, last) \
31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
H A Dmmu_context.h35 #define switch_mm(prev, next, tsk) \
37 if (prev != next) \
38 change_mm_context(&prev->context, &next->context, next->pgd); \
41 #define activate_mm(prev, next) \
43 change_mm_context(&prev->context, &next->context, next->pgd); \
/linux-4.4.14/arch/mips/include/asm/
H A Dswitch_to.h26 * @next: The task to begin executing.
27 * @next_ti: task_thread_info(next).
30 * the context of next. Returns prev.
33 struct task_struct *next, struct thread_info *next_ti);
62 next->thread.emulated_fp = 0; \
84 #define switch_to(prev, next, last) \
90 __restore_dsp(next); \
99 if (KSTK_STATUS(next) & ST0_CU2 && \
101 cop2_restore(next); \
107 write_c0_userlocal(task_thread_info(next)->tp_value); \
108 __restore_watch(next); \
109 (last) = resume(prev, next, task_thread_info(next)); \
H A Dmmu_context.h139 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
148 if ((cpu_context(cpu, next) ^ asid_cache(cpu)) & ASID_VERSION_MASK) switch_mm()
149 get_new_mmu_context(next, cpu); switch_mm()
150 write_c0_entryhi(cpu_asid(cpu, next)); switch_mm()
151 TLBMISS_HANDLER_SETUP_PGD(next->pgd); switch_mm()
158 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm()
179 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
188 get_new_mmu_context(next, cpu); activate_mm()
190 write_c0_entryhi(cpu_asid(cpu, next)); activate_mm()
191 TLBMISS_HANDLER_SETUP_PGD(next->pgd); activate_mm()
195 cpumask_set_cpu(cpu, mm_cpumask(next)); activate_mm()
217 /* will get a new context next time */ drop_mmu_context()
/linux-4.4.14/arch/blackfin/mm/
H A Dsram-alloc.c31 struct sram_piece *next; member in struct:sram_piece
78 per_cpu(free_l1_ssram_head, cpu).next = l1sram_init()
80 if (!per_cpu(free_l1_ssram_head, cpu).next) { l1sram_init()
85 per_cpu(free_l1_ssram_head, cpu).next->paddr = (void *)get_l1_scratch_start_cpu(cpu) + reserve; l1sram_init()
86 per_cpu(free_l1_ssram_head, cpu).next->size = L1_SCRATCH_LENGTH - reserve; l1sram_init()
87 per_cpu(free_l1_ssram_head, cpu).next->pid = 0; l1sram_init()
88 per_cpu(free_l1_ssram_head, cpu).next->next = NULL; l1sram_init()
90 per_cpu(used_l1_ssram_head, cpu).next = NULL; l1sram_init()
106 per_cpu(free_l1_data_A_sram_head, cpu).next = l1_data_sram_init()
108 if (!per_cpu(free_l1_data_A_sram_head, cpu).next) { l1_data_sram_init()
113 per_cpu(free_l1_data_A_sram_head, cpu).next->paddr = l1_data_sram_init()
115 per_cpu(free_l1_data_A_sram_head, cpu).next->size = l1_data_sram_init()
117 per_cpu(free_l1_data_A_sram_head, cpu).next->pid = 0; l1_data_sram_init()
118 per_cpu(free_l1_data_A_sram_head, cpu).next->next = NULL; l1_data_sram_init()
120 per_cpu(used_l1_data_A_sram_head, cpu).next = NULL; l1_data_sram_init()
124 per_cpu(free_l1_data_A_sram_head, cpu).next->size >> 10); l1_data_sram_init()
129 per_cpu(free_l1_data_B_sram_head, cpu).next = l1_data_sram_init()
131 if (!per_cpu(free_l1_data_B_sram_head, cpu).next) { l1_data_sram_init()
136 per_cpu(free_l1_data_B_sram_head, cpu).next->paddr = l1_data_sram_init()
138 per_cpu(free_l1_data_B_sram_head, cpu).next->size = l1_data_sram_init()
140 per_cpu(free_l1_data_B_sram_head, cpu).next->pid = 0; l1_data_sram_init()
141 per_cpu(free_l1_data_B_sram_head, cpu).next->next = NULL; l1_data_sram_init()
143 per_cpu(used_l1_data_B_sram_head, cpu).next = NULL; l1_data_sram_init()
147 per_cpu(free_l1_data_B_sram_head, cpu).next->size >> 10); l1_data_sram_init()
163 per_cpu(free_l1_inst_sram_head, cpu).next = l1_inst_sram_init()
165 if (!per_cpu(free_l1_inst_sram_head, cpu).next) { l1_inst_sram_init()
170 per_cpu(free_l1_inst_sram_head, cpu).next->paddr = l1_inst_sram_init()
172 per_cpu(free_l1_inst_sram_head, cpu).next->size = l1_inst_sram_init()
174 per_cpu(free_l1_inst_sram_head, cpu).next->pid = 0; l1_inst_sram_init()
175 per_cpu(free_l1_inst_sram_head, cpu).next->next = NULL; l1_inst_sram_init()
177 per_cpu(used_l1_inst_sram_head, cpu).next = NULL; l1_inst_sram_init()
181 per_cpu(free_l1_inst_sram_head, cpu).next->size >> 10); l1_inst_sram_init()
228 free_l2_sram_head.next = l2_sram_init()
230 if (!free_l2_sram_head.next) { l2_sram_init()
235 free_l2_sram_head.next->paddr = l2_sram_init()
237 free_l2_sram_head.next->size = l2_sram_init()
239 free_l2_sram_head.next->pid = 0; l2_sram_init()
240 free_l2_sram_head.next->next = NULL; l2_sram_init()
242 used_l2_sram_head.next = NULL; l2_sram_init()
246 free_l2_sram_head.next->size >> 10); l2_sram_init()
280 pslot = pfree_head->next; _sram_alloc()
286 pslot = pslot->next; _sram_alloc()
293 plast->next = pslot->next; _sram_alloc()
310 pslot = pused_head->next; _sram_alloc()
316 pslot = pslot->next; _sram_alloc()
319 pavail->next = pslot; _sram_alloc()
320 plast->next = pavail; _sram_alloc()
335 pmax = pslot = pfree_head->next; _sram_alloc_max()
341 pslot = pslot->next; _sram_alloc_max()
363 pslot = pused_head->next; _sram_free()
369 pslot = pslot->next; _sram_free()
375 plast->next = pslot->next; _sram_free()
380 pslot = pfree_head->next; _sram_free()
385 pslot = pslot->next; _sram_free()
392 pavail->next = plast->next; _sram_free()
393 plast->next = pavail; _sram_free()
399 plast->next = pslot->next; _sram_free()
746 for (tmp = &mm->context.sram_list; *tmp; tmp = &(*tmp)->next) sram_free_with_lsl()
750 *tmp = lsl->next; sram_free_with_lsl()
791 lsl->next = mm->context.sram_list; sram_alloc_with_lsl()
816 pslot = pused_head->next; _sram_proc_show()
823 pslot = pslot->next; _sram_proc_show()
826 pslot = pfree_head->next; _sram_proc_show()
833 pslot = pslot->next; _sram_proc_show()
/linux-4.4.14/net/netlabel/
H A Dnetlabel_addrlist.h78 i = i->next; __af4list_valid()
90 i = rcu_dereference(i->next); __af4list_valid_rcu()
97 for (iter = __af4list_valid((head)->next, head); \
99 iter = __af4list_valid(iter->list.next, head))
102 for (iter = __af4list_valid_rcu((head)->next, head); \
104 iter = __af4list_valid_rcu(iter->list.next, head))
107 for (iter = __af4list_valid((head)->next, head), \
108 tmp = __af4list_valid(iter->list.next, head); \
110 iter = tmp, tmp = __af4list_valid(iter->list.next, head))
145 i = i->next; __af6list_valid()
157 i = rcu_dereference(i->next); __af6list_valid_rcu()
164 for (iter = __af6list_valid((head)->next, head); \
166 iter = __af6list_valid(iter->list.next, head))
169 for (iter = __af6list_valid_rcu((head)->next, head); \
171 iter = __af6list_valid_rcu(iter->list.next, head))
174 for (iter = __af6list_valid((head)->next, head), \
175 tmp = __af6list_valid(iter->list.next, head); \
177 iter = tmp, tmp = __af6list_valid(iter->list.next, head))
/linux-4.4.14/arch/sparc/include/asm/
H A Dswitch_to_64.h6 #define prepare_arch_switch(next) \
14 * for l0/l1. It will use one for 'next' and the other to hold
15 * the output value of 'last'. 'next' is not referenced again
20 #define switch_to(prev, next, last) \
25 : : "r" (task_thread_info(next)->current_ds));\
27 task_thread_info(next); \
58 : "0" (task_thread_info(next)), \
H A Dswitch_to_32.h27 #define SWITCH_DO_LAZY_FPU(next) /* */
37 #define prepare_arch_switch(next) do { \
57 #define switch_to(prev, next, last) do { \
59 SWITCH_DO_LAZY_FPU(next); \
60 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next->active_mm)); \
92 "r" (task_thread_info(next)), \
H A Dcpudata_32.h24 int next; member in struct:__anon2689
/linux-4.4.14/arch/m32r/include/asm/
H A Dswitch_to.h13 * switch_to(prev, next) should switch from task `prev' to `next'
14 * `prev' will never be the same as `next'.
16 * `next' and `prev' should be struct task_struct, but it isn't always defined
28 #define switch_to(prev, next, last) do { \
45 "r" (&(prev->thread.sp)), "r" (&(next->thread.sp)), \
46 "r" (&(prev->thread.lr)), "r" (&(next->thread.lr)) \
H A Dmmu_context.h122 struct mm_struct *next, struct task_struct *tsk) switch_mm()
128 if (prev != next) { switch_mm()
130 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm()
132 /* Set MPTB = next->pgd */ switch_mm()
133 *(volatile unsigned long *)MPTB = (unsigned long)next->pgd; switch_mm()
134 activate_context(next); switch_mm()
138 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) switch_mm()
139 activate_context(next); switch_mm()
145 #define activate_mm(prev, next) \
146 switch_mm((prev), (next), NULL)
155 #define switch_mm(prev,next,tsk) do { } while (0)
157 #define activate_mm(prev,next) do { } while (0)
121 switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
/linux-4.4.14/drivers/gpu/drm/radeon/
H A Dmkregtable.c36 * sometimes we already know the next/prev entries and we can
42 struct list_head *next, *prev; member in struct:list_head
52 list->next = list; INIT_LIST_HEAD()
60 * the prev/next entries already!
64 struct list_head *prev, struct list_head *next) __list_add()
66 next->prev = new; __list_add()
67 new->next = next; __list_add()
69 prev->next = new; __list_add()
73 struct list_head *prev, struct list_head *next);
86 __list_add(new, head, head->next); list_add()
103 * Delete a list entry by making the prev/next entries
107 * the prev/next entries already!
109 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument
111 next->prev = prev; __list_del()
112 prev->next = next; __list_del()
124 __list_del(entry->prev, entry->next); list_del()
125 entry->next = (void *)0xDEADBEEF; list_del()
141 new->next = old->next; list_replace()
142 new->next->prev = new; list_replace()
144 new->prev->next = new; list_replace()
160 __list_del(entry->prev, entry->next); list_del_init()
171 __list_del(list->prev, list->next); list_move()
183 __list_del(list->prev, list->next); list_move_tail()
195 return list->next == head; list_is_last()
204 return head->next == head; list_empty()
213 * in the process of modifying either member (next or prev)
222 struct list_head *next = head->next; list_empty_careful() local
223 return (next == head) && (next == head->prev); list_empty_careful()
232 return !list_empty(head) && (head->next == head->prev); list_is_singular()
239 struct list_head *new_first = entry->next; __list_cut_position()
240 list->next = head->next; __list_cut_position()
241 list->next->prev = list; __list_cut_position()
243 entry->next = list; __list_cut_position()
244 head->next = new_first; __list_cut_position()
268 if (list_is_singular(head) && (head->next != entry && head != entry)) list_cut_position()
277 struct list_head *prev, struct list_head *next) __list_splice()
279 struct list_head *first = list->next; __list_splice()
283 prev->next = first; __list_splice()
285 last->next = next; __list_splice()
286 next->prev = last; __list_splice()
298 __list_splice(list, head, head->next); list_splice()
324 __list_splice(list, head, head->next); list_splice_init()
364 list_entry((ptr)->next, type, member)
372 for (pos = (head)->next; prefetch(pos->next), pos != (head); \
373 pos = pos->next)
391 for (pos = (head)->next, n = pos->next; pos != (head); \
392 pos = n, n = pos->next)
412 for (pos = list_entry((head)->next, typeof(*pos), member); \
414 pos = list_entry(pos->member.next, typeof(*pos), member))
448 for (pos = list_entry(pos->member.next, typeof(*pos), member); \
449 prefetch(pos->member.next), &pos->member != (head); \
450 pos = list_entry(pos->member.next, typeof(*pos), member))
475 for (; prefetch(pos->member.next), &pos->member != (head); \
476 pos = list_entry(pos->member.next, typeof(*pos), member))
486 for (pos = list_entry((head)->next, typeof(*pos), member), \
487 n = list_entry(pos->member.next, typeof(*pos), member); \
489 pos = n, n = list_entry(n->member.next, typeof(*n), member))
502 for (pos = list_entry(pos->member.next, typeof(*pos), member), \
503 n = list_entry(pos->member.next, typeof(*pos), member); \
505 pos = n, n = list_entry(n->member.next, typeof(*n), member))
518 for (n = list_entry(pos->member.next, typeof(*pos), member); \
520 pos = n, n = list_entry(n->member.next, typeof(*n), member))
63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
276 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
H A Dradeon_mem.c16 * The above copyright notice and this permission notice (including the next
54 newblock->next = p->next; split_block()
56 p->next->prev = newblock; split_block()
57 p->next = newblock; split_block()
71 newblock->next = p->next; split_block()
73 p->next->prev = newblock; split_block()
74 p->next = newblock; split_block()
117 if (p->next->file_priv == NULL) { free_block()
118 struct mem_block *q = p->next; free_block()
120 p->next = q->next; free_block()
121 p->next->prev = p; free_block()
128 q->next = p->next; free_block()
129 q->next->prev = q; free_block()
152 blocks->next = blocks->prev = *heap; init_heap()
155 (*heap)->next = (*heap)->prev = blocks; init_heap()
165 if (!heap || !heap->next) radeon_mem_release()
177 while (p->file_priv == NULL && p->next->file_priv == NULL) { list_for_each()
178 struct mem_block *q = p->next; list_for_each()
180 p->next = q->next; list_for_each()
181 p->next->prev = p; list_for_each()
196 for (p = (*heap)->next; p != *heap;) { radeon_mem_takedown()
198 p = p->next; radeon_mem_takedown()
/linux-4.4.14/security/selinux/ss/
H A Dhashtab.c54 cur = cur->next; hashtab_insert()
66 newnode->next = prev->next; hashtab_insert()
67 prev->next = newnode; hashtab_insert()
69 newnode->next = h->htable[hvalue]; hashtab_insert()
88 cur = cur->next; hashtab_search()
108 cur = cur->next; hashtab_destroy()
137 cur = cur->next; hashtab_map()
158 cur = cur->next; hashtab_stat()
H A Debitmap.c39 n1 = n1->next; ebitmap_cmp()
40 n2 = n2->next; ebitmap_cmp()
64 new->next = NULL; ebitmap_cpy()
66 prev->next = new; ebitmap_cpy()
70 n = n->next; ebitmap_cpy()
120 e_iter = e_iter->next; ebitmap_netlbl_export()
172 e_prev->next = e_iter; ebitmap_netlbl_import()
180 /* next */ ebitmap_netlbl_import()
211 n1 = n1->next; ebitmap_contains()
229 n1 = n1->next; ebitmap_contains()
230 n2 = n2->next; ebitmap_contains()
250 n = n->next; ebitmap_get_bit()
276 if (!n->next) { ebitmap_set_bit()
288 prev->next = n->next; ebitmap_set_bit()
290 e->node = n->next; ebitmap_set_bit()
296 n = n->next; ebitmap_set_bit()
314 new->next = prev->next; ebitmap_set_bit()
315 prev->next = new; ebitmap_set_bit()
317 new->next = e->node; ebitmap_set_bit()
334 n = n->next; ebitmap_destroy()
410 n->next = tmp; ebitmap_read()
496 /* set up for the next node */ ebitmap_for_each_positive_bit()
H A Dsidtab.c48 cur = cur->next; sidtab_insert()
69 newnode->next = prev->next; sidtab_insert()
71 prev->next = newnode; sidtab_insert()
73 newnode->next = s->htable[hvalue]; sidtab_insert()
96 cur = cur->next; sidtab_search_core()
107 cur = cur->next; sidtab_search_core()
143 cur = cur->next; sidtab_map()
174 cur = cur->next; sidtab_search_context()
254 cur = cur->next; sidtab_hash_eval()
279 cur = cur->next; sidtab_destroy()
H A Dconstraint.h53 struct constraint_expr *next; /* next expression */ member in struct:constraint_expr
59 struct constraint_node *next; /* next constraint */ member in struct:constraint_node
H A Dsidtab.h15 struct sidtab_node *next; member in struct:sidtab_node
27 unsigned int next_sid; /* next SID to allocate */
H A Dconditional.c33 for (cur = expr; cur; cur = cur->next) { cond_evaluate_expr()
101 for (cur = node->true_list; cur; cur = cur->next) { evaluate_cond_node()
108 for (cur = node->false_list; cur; cur = cur->next) { evaluate_cond_node()
135 struct cond_av_list *cur, *next; cond_av_list_destroy() local
136 for (cur = list; cur; cur = next) { cond_av_list_destroy()
137 next = cur->next; cond_av_list_destroy()
148 next_expr = cur_expr->next; cond_node_destroy()
158 struct cond_node *next, *cur; cond_list_destroy() local
163 for (cur = list; cur; cur = next) { cond_list_destroy()
164 next = cur->next; cond_list_destroy()
306 for (cur = other; cur; cur = cur->next) { cond_insertf()
342 data->tail->next = list; cond_insertf()
437 last->next = expr; cond_read_node()
483 last->next = node; cond_read_list()
534 for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) cond_write_av_list()
545 for (cur_list = list; cur_list != NULL; cur_list = cur_list->next) { cond_write_av_list()
567 for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) cond_write_node()
575 for (cur_expr = node->expr; cur_expr != NULL; cur_expr = cur_expr->next) { cond_write_node()
601 for (cur = list; cur != NULL; cur = cur->next) cond_write_list()
608 for (cur = list; cur != NULL; cur = cur->next) { cond_write_list()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dgddr3.c75 switch (ram->next->bios.timing_ver) { nvkm_gddr3_calc()
77 CWL = ram->next->bios.timing_10_CWL; nvkm_gddr3_calc()
78 CL = ram->next->bios.timing_10_CL; nvkm_gddr3_calc()
79 WR = ram->next->bios.timing_10_WR; nvkm_gddr3_calc()
80 DLL = !ram->next->bios.ramcfg_DLLoff; nvkm_gddr3_calc()
81 ODT = ram->next->bios.timing_10_ODT; nvkm_gddr3_calc()
82 RON = ram->next->bios.ramcfg_RON; nvkm_gddr3_calc()
85 CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; nvkm_gddr3_calc()
86 CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; nvkm_gddr3_calc()
87 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_gddr3_calc()
96 if (ram->next->bios.timing_ver == 0x20 || nvkm_gddr3_calc()
97 ram->next->bios.ramcfg_timing == 0xff) { nvkm_gddr3_calc()
H A Dgddr5.c41 xd = !ram->next->bios.ramcfg_DLLoff; nvkm_gddr5_calc()
43 switch (ram->next->bios.ramcfg_ver) { nvkm_gddr5_calc()
45 pd = ram->next->bios.ramcfg_11_01_80; nvkm_gddr5_calc()
46 lf = ram->next->bios.ramcfg_11_01_40; nvkm_gddr5_calc()
47 vh = ram->next->bios.ramcfg_11_02_10; nvkm_gddr5_calc()
48 vr = ram->next->bios.ramcfg_11_02_04; nvkm_gddr5_calc()
49 vo = ram->next->bios.ramcfg_11_06; nvkm_gddr5_calc()
50 l3 = !ram->next->bios.ramcfg_11_07_02; nvkm_gddr5_calc()
56 switch (ram->next->bios.timing_ver) { nvkm_gddr5_calc()
58 WL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; nvkm_gddr5_calc()
59 CL = (ram->next->bios.timing[1] & 0x0000001f); nvkm_gddr5_calc()
60 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_gddr5_calc()
61 at[0] = ram->next->bios.timing_20_2e_c0; nvkm_gddr5_calc()
62 at[1] = ram->next->bios.timing_20_2e_30; nvkm_gddr5_calc()
63 dt = ram->next->bios.timing_20_2e_03; nvkm_gddr5_calc()
64 ds = ram->next->bios.timing_20_2f_03; nvkm_gddr5_calc()
H A Dsddr3.c73 DLL = !ram->next->bios.ramcfg_DLLoff; nvkm_sddr3_calc()
75 switch (ram->next->bios.timing_ver) { nvkm_sddr3_calc()
77 if (ram->next->bios.timing_hdr < 0x17) { nvkm_sddr3_calc()
81 CWL = ram->next->bios.timing_10_CWL; nvkm_sddr3_calc()
82 CL = ram->next->bios.timing_10_CL; nvkm_sddr3_calc()
83 WR = ram->next->bios.timing_10_WR; nvkm_sddr3_calc()
84 ODT = ram->next->bios.timing_10_ODT; nvkm_sddr3_calc()
87 CWL = (ram->next->bios.timing[1] & 0x00000f80) >> 7; nvkm_sddr3_calc()
88 CL = (ram->next->bios.timing[1] & 0x0000001f) >> 0; nvkm_sddr3_calc()
89 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_sddr3_calc()
H A Dramgk104.c210 struct nvkm_ram_data *next = ram->base.next; r1373f4_fini() local
211 u8 v0 = next->bios.ramcfg_11_03_c0; r1373f4_fini()
212 u8 v1 = next->bios.ramcfg_11_03_30; r1373f4_fini()
243 u32 next = (prev & ~mask) | data; gk104_ram_nuts() local
244 nvkm_memx_wr32(fuc->memx, addr, next); gk104_ram_nuts()
255 struct nvkm_ram_data *next = ram->base.next; gk104_ram_calc_gddr5() local
256 int vc = !next->bios.ramcfg_11_02_08; gk104_ram_calc_gddr5()
257 int mv = !next->bios.ramcfg_11_02_04; gk104_ram_calc_gddr5()
309 switch (next->bios.ramcfg_11_03_c0) { gk104_ram_calc_gddr5()
316 switch (next->bios.ramcfg_11_03_30) { gk104_ram_calc_gddr5()
324 if (next->bios.ramcfg_11_02_80) gk104_ram_calc_gddr5()
326 if (next->bios.ramcfg_11_02_40) gk104_ram_calc_gddr5()
328 if (next->bios.ramcfg_11_07_10) gk104_ram_calc_gddr5()
330 if (next->bios.ramcfg_11_07_08) gk104_ram_calc_gddr5()
365 if (next->bios.ramcfg_11_02_40 || gk104_ram_calc_gddr5()
366 next->bios.ramcfg_11_07_10) { gk104_ram_calc_gddr5()
389 if (next->bios.ramcfg_11_07_40) gk104_ram_calc_gddr5()
393 ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c); gk104_ram_calc_gddr5()
394 ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_gddr5()
395 ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_gddr5()
397 if (!next->bios.ramcfg_11_07_08 && !next->bios.ramcfg_11_07_04) { gk104_ram_calc_gddr5()
398 ram_wr32(fuc, 0x10f698, 0x01010101 * next->bios.ramcfg_11_04); gk104_ram_calc_gddr5()
399 ram_wr32(fuc, 0x10f69c, 0x01010101 * next->bios.ramcfg_11_04); gk104_ram_calc_gddr5()
401 if (!next->bios.ramcfg_11_07_08) { gk104_ram_calc_gddr5()
407 u32 data = 0x01000100 * next->bios.ramcfg_11_04; gk104_ram_calc_gddr5()
412 if (ram->mode == 2 && next->bios.ramcfg_11_08_10) gk104_ram_calc_gddr5()
420 if (!next->bios.ramcfg_11_02_80) gk104_ram_calc_gddr5()
422 if (!next->bios.ramcfg_11_02_40) gk104_ram_calc_gddr5()
424 if (!next->bios.ramcfg_11_07_10) gk104_ram_calc_gddr5()
426 if (!next->bios.ramcfg_11_07_08) gk104_ram_calc_gddr5()
432 if (next->bios.ramcfg_11_01_08) gk104_ram_calc_gddr5()
443 if (next->bios.ramcfg_11_08_01) gk104_ram_calc_gddr5()
450 if (next->bios.ramcfg_11_08_08) gk104_ram_calc_gddr5()
452 if (next->bios.ramcfg_11_08_04) gk104_ram_calc_gddr5()
454 if (next->bios.ramcfg_11_08_02) gk104_ram_calc_gddr5()
459 ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]); gk104_ram_calc_gddr5()
460 ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]); gk104_ram_calc_gddr5()
461 ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]); gk104_ram_calc_gddr5()
462 ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]); gk104_ram_calc_gddr5()
463 ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]); gk104_ram_calc_gddr5()
464 ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]); gk104_ram_calc_gddr5()
465 ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]); gk104_ram_calc_gddr5()
466 ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]); gk104_ram_calc_gddr5()
467 ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]); gk104_ram_calc_gddr5()
468 ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]); gk104_ram_calc_gddr5()
469 ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]); gk104_ram_calc_gddr5()
473 if (next->bios.ramcfg_11_08_20) gk104_ram_calc_gddr5()
481 data |= next->bios.ramcfg_11_02_03 << 8; gk104_ram_calc_gddr5()
485 if (next->bios.ramcfg_11_01_10) gk104_ram_calc_gddr5()
493 data |= next->bios.timing_20_30_07 << 28; gk104_ram_calc_gddr5()
497 if (next->bios.ramcfg_11_01_01) gk104_ram_calc_gddr5()
505 data |= next->bios.timing_20_30_07 << 28; gk104_ram_calc_gddr5()
509 if (next->bios.ramcfg_11_01_02) gk104_ram_calc_gddr5()
517 if (!next->bios.ramcfg_11_01_04) gk104_ram_calc_gddr5()
519 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_gddr5()
524 if (next->bios.ramcfg_11_03_f0) { gk104_ram_calc_gddr5()
525 if (next->bios.rammap_11_08_0c) { gk104_ram_calc_gddr5()
526 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_gddr5()
539 ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f); gk104_ram_calc_gddr5()
543 data |= next->bios.ramcfg_11_02_03; gk104_ram_calc_gddr5()
547 if (next->bios.ramcfg_11_01_10) gk104_ram_calc_gddr5()
558 data = next->bios.timing_20_30_07 << 8; gk104_ram_calc_gddr5()
559 if (next->bios.ramcfg_11_01_01) gk104_ram_calc_gddr5()
563 ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4); gk104_ram_calc_gddr5()
564 data = (next->bios.timing[10] & 0x7f000000) >> 24; gk104_ram_calc_gddr5()
565 if (data < next->bios.timing_20_2c_1fc0) gk104_ram_calc_gddr5()
566 data = next->bios.timing_20_2c_1fc0; gk104_ram_calc_gddr5()
568 ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16); gk104_ram_calc_gddr5()
570 ram_mask(fuc, 0x10fec4, 0x041e0f07, next->bios.timing_20_31_0800 << 26 | gk104_ram_calc_gddr5()
571 next->bios.timing_20_31_0780 << 17 | gk104_ram_calc_gddr5()
572 next->bios.timing_20_31_0078 << 8 | gk104_ram_calc_gddr5()
573 next->bios.timing_20_31_0007); gk104_ram_calc_gddr5()
574 ram_mask(fuc, 0x10fec8, 0x00000027, next->bios.timing_20_31_8000 << 5 | gk104_ram_calc_gddr5()
575 next->bios.timing_20_31_7000); gk104_ram_calc_gddr5()
583 if (next->bios.ramcfg_11_08_10 && (ram->mode == 2) /*XXX*/) { gk104_ram_calc_gddr5()
616 if (!next->bios.ramcfg_11_07_08) { gk104_ram_calc_gddr5()
617 if (!next->bios.ramcfg_11_07_04) gk104_ram_calc_gddr5()
631 if (!next->bios.ramcfg_11_07_08) { gk104_ram_calc_gddr5()
633 if ( next->bios.ramcfg_11_07_04) gk104_ram_calc_gddr5()
635 if (!next->bios.rammap_11_08_10) gk104_ram_calc_gddr5()
657 if (next->bios.ramcfg_11_07_02) gk104_ram_calc_gddr5()
663 if (next->bios.rammap_11_08_01) gk104_ram_calc_gddr5()
702 struct nvkm_ram_data *next = ram->base.next; gk104_ram_calc_sddr3() local
703 int vc = !next->bios.ramcfg_11_02_08; gk104_ram_calc_sddr3()
704 int mv = !next->bios.ramcfg_11_02_04; gk104_ram_calc_sddr3()
720 if (next->bios.ramcfg_11_03_f0) gk104_ram_calc_sddr3()
725 if (next->bios.ramcfg_DLLoff) gk104_ram_calc_sddr3()
749 switch (next->bios.ramcfg_11_03_c0) { gk104_ram_calc_sddr3()
756 switch (next->bios.ramcfg_11_03_30) { gk104_ram_calc_sddr3()
764 if (next->bios.ramcfg_11_02_80) gk104_ram_calc_sddr3()
766 if (next->bios.ramcfg_11_02_40) gk104_ram_calc_sddr3()
768 if (next->bios.ramcfg_11_07_10) gk104_ram_calc_sddr3()
770 if (next->bios.ramcfg_11_07_08) gk104_ram_calc_sddr3()
780 data |= next->bios.ramcfg_11_03_30 << 16; gk104_ram_calc_sddr3()
812 if (next->bios.ramcfg_11_02_40 || gk104_ram_calc_sddr3()
813 next->bios.ramcfg_11_07_10) { gk104_ram_calc_sddr3()
819 if (next->bios.ramcfg_11_07_40) gk104_ram_calc_sddr3()
823 ram_wr32(fuc, 0x10f65c, 0x00000011 * next->bios.rammap_11_11_0c); gk104_ram_calc_sddr3()
824 ram_wr32(fuc, 0x10f6b8, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_sddr3()
825 ram_wr32(fuc, 0x10f6bc, 0x01010101 * next->bios.ramcfg_11_09); gk104_ram_calc_sddr3()
829 if (!next->bios.ramcfg_11_02_80) gk104_ram_calc_sddr3()
831 if (!next->bios.ramcfg_11_02_40) gk104_ram_calc_sddr3()
833 if (!next->bios.ramcfg_11_07_10) gk104_ram_calc_sddr3()
835 if (!next->bios.ramcfg_11_07_08) gk104_ram_calc_sddr3()
842 if (next->bios.ramcfg_11_08_01) gk104_ram_calc_sddr3()
849 ram_mask(fuc, 0x10f248, 0xffffffff, next->bios.timing[10]); gk104_ram_calc_sddr3()
850 ram_mask(fuc, 0x10f290, 0xffffffff, next->bios.timing[0]); gk104_ram_calc_sddr3()
851 ram_mask(fuc, 0x10f294, 0xffffffff, next->bios.timing[1]); gk104_ram_calc_sddr3()
852 ram_mask(fuc, 0x10f298, 0xffffffff, next->bios.timing[2]); gk104_ram_calc_sddr3()
853 ram_mask(fuc, 0x10f29c, 0xffffffff, next->bios.timing[3]); gk104_ram_calc_sddr3()
854 ram_mask(fuc, 0x10f2a0, 0xffffffff, next->bios.timing[4]); gk104_ram_calc_sddr3()
855 ram_mask(fuc, 0x10f2a4, 0xffffffff, next->bios.timing[5]); gk104_ram_calc_sddr3()
856 ram_mask(fuc, 0x10f2a8, 0xffffffff, next->bios.timing[6]); gk104_ram_calc_sddr3()
857 ram_mask(fuc, 0x10f2ac, 0xffffffff, next->bios.timing[7]); gk104_ram_calc_sddr3()
858 ram_mask(fuc, 0x10f2cc, 0xffffffff, next->bios.timing[8]); gk104_ram_calc_sddr3()
859 ram_mask(fuc, 0x10f2e8, 0xffffffff, next->bios.timing[9]); gk104_ram_calc_sddr3()
863 if (!next->bios.ramcfg_11_01_04) gk104_ram_calc_sddr3()
865 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_sddr3()
870 if (next->bios.ramcfg_11_03_f0) { gk104_ram_calc_sddr3()
871 if (next->bios.rammap_11_08_0c) { gk104_ram_calc_sddr3()
872 if (!next->bios.ramcfg_11_07_80) gk104_ram_calc_sddr3()
886 ram_wr32(fuc, 0x10f870, 0x11111111 * next->bios.ramcfg_11_03_0f); gk104_ram_calc_sddr3()
888 ram_mask(fuc, 0x10f250, 0x000003f0, next->bios.timing_20_2c_003f << 4); gk104_ram_calc_sddr3()
890 data = (next->bios.timing[10] & 0x7f000000) >> 24; gk104_ram_calc_sddr3()
891 if (data < next->bios.timing_20_2c_1fc0) gk104_ram_calc_sddr3()
892 data = next->bios.timing_20_2c_1fc0; gk104_ram_calc_sddr3()
895 ram_mask(fuc, 0x10f224, 0x001f0000, next->bios.timing_20_30_f8 << 16); gk104_ram_calc_sddr3()
905 if (!next->bios.ramcfg_DLLoff) { gk104_ram_calc_sddr3()
915 if (!next->bios.ramcfg_DLLoff) { gk104_ram_calc_sddr3()
941 if (next->bios.rammap_11_08_01) gk104_ram_calc_sddr3()
1035 gk104_ram_calc_xits(struct gk104_ram *ram, struct nvkm_ram_data *next) gk104_ram_calc_xits() argument
1046 ram->mode = (next->freq > fuc->refpll.vco1.max_freq) ? 2 : 1; gk104_ram_calc_xits()
1057 refclk = next->freq; gk104_ram_calc_xits()
1059 ret = gk104_pll_calc_hiclk(next->freq, subdev->device->crystal, gk104_ram_calc_xits()
1068 " (refclock: %i kHz)\n", next->freq, ret); gk104_ram_calc_xits()
1084 ram->base.freq = next->freq; gk104_ram_calc_xits()
1090 ret = gk104_ram_calc_sddr3(ram, next->freq); gk104_ram_calc_xits()
1095 ret = gk104_ram_calc_gddr5(ram, next->freq); gk104_ram_calc_xits()
1114 if (ram->base.next == NULL) { gk104_ram_calc()
1137 ram->base.next = &ram->base.target; gk104_ram_calc()
1139 ram->base.next = &ram->base.xition; gk104_ram_calc()
1141 BUG_ON(ram->base.next != &ram->base.xition); gk104_ram_calc()
1142 ram->base.next = &ram->base.target; gk104_ram_calc()
1145 return gk104_ram_calc_xits(ram, ram->base.next); gk104_ram_calc()
1230 struct nvkm_ram_data *next = ram->base.next; gk104_ram_prog() local
1234 return (ram->base.next == &ram->base.xition); gk104_ram_prog()
1239 gk104_ram_prog_0(ram, next->freq); gk104_ram_prog()
1241 return (ram->base.next == &ram->base.xition); gk104_ram_prog()
1248 ram->base.next = NULL; gk104_ram_tidy()
H A Dsddr2.c64 switch (ram->next->bios.timing_ver) { nvkm_sddr2_calc()
66 CL = ram->next->bios.timing_10_CL; nvkm_sddr2_calc()
67 WR = ram->next->bios.timing_10_WR; nvkm_sddr2_calc()
68 DLL = !ram->next->bios.ramcfg_DLLoff; nvkm_sddr2_calc()
69 ODT = ram->next->bios.timing_10_ODT & 3; nvkm_sddr2_calc()
72 CL = (ram->next->bios.timing[1] & 0x0000001f); nvkm_sddr2_calc()
73 WR = (ram->next->bios.timing[2] & 0x007f0000) >> 16; nvkm_sddr2_calc()
79 if (ram->next->bios.timing_ver == 0x20 || nvkm_sddr2_calc()
80 ram->next->bios.ramcfg_timing == 0xff) { nvkm_sddr2_calc()
H A Dramnv50.c97 unkt3b = 0x19 + ram->base.next->bios.rammap_00_16_40; nv50_ram_timing_calc()
99 ram->base.next->bios.rammap_00_16_40) << 16 | nv50_ram_timing_calc()
227 struct nvkm_ram_data *next; nv50_ram_calc() local
235 next = &ram->base.target; nv50_ram_calc()
236 next->freq = freq; nv50_ram_calc()
237 ram->base.next = next; nv50_ram_calc()
251 nvbios_rammapEp_from_perf(bios, data, hdr, &next->bios); nv50_ram_calc()
261 &next->bios); nv50_ram_calc()
268 if (next->bios.ramcfg_timing != 0xff) { nv50_ram_calc()
269 data = nvbios_timingEp(bios, next->bios.ramcfg_timing, nv50_ram_calc()
270 &ver, &hdr, &cnt, &len, &next->bios); nv50_ram_calc()
305 if (subdev->device->chipset <= 0x96 && !next->bios.ramcfg_00_03_02) nv50_ram_calc()
319 if (next->bios.timing_10_ODT) nv50_ram_calc()
355 next->bios.rammap_00_16_40 << 14); nv50_ram_calc()
363 nv50_ram_gpio(hwsq, 0x18, !next->bios.ramcfg_FBVDDQ); nv50_ram_calc()
400 if (!next->bios.ramcfg_00_03_02) nv50_ram_calc()
402 ram_mask(hwsq, 0x100200, 0x00001000, !next->bios.ramcfg_00_04_02 << 12); nv50_ram_calc()
413 if (!next->bios.ramcfg_00_03_08) nv50_ram_calc()
415 if (!next->bios.ramcfg_FBVDDQ) nv50_ram_calc()
417 if ( next->bios.ramcfg_00_04_04) nv50_ram_calc()
422 if (!next->bios.ramcfg_00_03_08) nv50_ram_calc()
426 if ( next->bios.ramcfg_00_03_01) nv50_ram_calc()
428 if ( next->bios.ramcfg_00_03_02) nv50_ram_calc()
430 if (!next->bios.ramcfg_00_03_08) nv50_ram_calc()
432 if ( next->bios.ramcfg_00_04_04) nv50_ram_calc()
434 if ( next->bios.ramcfg_00_04_20) nv50_ram_calc()
444 if (next->bios.rammap_00_16_20) { nv50_ram_calc()
445 ram_wr32(hwsq, 0x1005a0, next->bios.ramcfg_00_07 << 16 | nv50_ram_calc()
446 next->bios.ramcfg_00_06 << 8 | nv50_ram_calc()
447 next->bios.ramcfg_00_05); nv50_ram_calc()
448 ram_wr32(hwsq, 0x1005a4, next->bios.ramcfg_00_09 << 8 | nv50_ram_calc()
449 next->bios.ramcfg_00_08); nv50_ram_calc()
456 if (!next->bios.timing_10_ODT) nv50_ram_calc()
460 if (!next->bios.ramcfg_DLLoff) nv50_ram_calc()
468 if (next->bios.rammap_00_17_02) nv50_ram_calc()
470 if (!next->bios.rammap_00_16_40) nv50_ram_calc()
472 if (next->bios.ramcfg_00_03_02) nv50_ram_calc()
474 if (subdev->device->chipset <= 0x96 && next->bios.ramcfg_00_03_02) nv50_ram_calc()
/linux-4.4.14/arch/c6x/include/asm/
H A Dswitch_to.h21 struct thread_struct *next,
24 #define switch_to(prev, next, last) \
28 &(next)->thread, (prev)); \
/linux-4.4.14/mm/
H A Dpagewalk.c31 unsigned long next; walk_pmd_range() local
37 next = pmd_addr_end(addr, end); walk_pmd_range()
40 err = walk->pte_hole(addr, next, walk); walk_pmd_range()
50 err = walk->pmd_entry(pmd, addr, next, walk); walk_pmd_range()
64 err = walk_pte_range(pmd, addr, next, walk); walk_pmd_range()
67 } while (pmd++, addr = next, addr != end); walk_pmd_range()
76 unsigned long next; walk_pud_range() local
81 next = pud_addr_end(addr, end); walk_pud_range()
84 err = walk->pte_hole(addr, next, walk); walk_pud_range()
90 err = walk_pmd_range(pud, addr, next, walk); walk_pud_range()
93 } while (pud++, addr = next, addr != end); walk_pud_range()
102 unsigned long next; walk_pgd_range() local
107 next = pgd_addr_end(addr, end); walk_pgd_range()
110 err = walk->pte_hole(addr, next, walk); walk_pgd_range()
116 err = walk_pud_range(pgd, addr, next, walk); walk_pgd_range()
119 } while (pgd++, addr = next, addr != end); walk_pgd_range()
137 unsigned long next; walk_hugetlb_range() local
143 next = hugetlb_entry_end(h, addr, end); walk_hugetlb_range()
146 err = walk->hugetlb_entry(pte, hmask, addr, next, walk); walk_hugetlb_range()
149 } while (addr = next, addr != end); walk_hugetlb_range()
243 unsigned long next; walk_page_range() local
258 next = end; walk_page_range()
261 next = min(end, vma->vm_start); walk_page_range()
264 next = min(end, vma->vm_end); walk_page_range()
267 err = walk_page_test(start, next, walk); walk_page_range()
281 err = __walk_page_range(start, next, walk); walk_page_range()
284 } while (start = next, start < end); walk_page_range()
H A Dslob.c22 * sufficient free blocks (using a next-fit-like approach) followed by
78 * or offset of next block if -ve (in SLOB_UNITs).
80 * Free blocks of size 1 unit simply contain the offset of the next block.
82 * memory, and the offset of the next free block in the second SLOB_UNIT.
143 * Encode the given size and next info into a free slob block s.
145 static void set_slob(slob_t *s, slobidx_t size, slob_t *next) set_slob() argument
148 slobidx_t offset = next - base; set_slob()
168 * Return the next free slob block pointer after this one.
173 slobidx_t next; slob_next() local
176 next = -s[0].units; slob_next()
178 next = s[1].units; slob_next()
179 return base+next; slob_next()
230 slob_t *next; slob_page_alloc() local
233 next = slob_next(cur); slob_page_alloc()
234 set_slob(aligned, avail - delta, next); slob_page_alloc()
241 next = slob_next(cur); slob_page_alloc()
244 set_slob(prev, slob_units(prev), next); slob_page_alloc()
246 sp->freelist = next; slob_page_alloc()
252 set_slob(cur + units, avail - units, next); slob_page_alloc()
305 * search time by starting our next search here. (see list_for_each_entry()
308 slob_list->next != prev->next) list_for_each_entry()
309 list_move_tail(slob_list, prev->next); list_for_each_entry()
343 slob_t *prev, *next, *b = (slob_t *)block; slob_free() local
400 next = slob_next(prev); slob_free()
401 while (b > next) { slob_free()
402 prev = next; slob_free()
403 next = slob_next(prev); slob_free()
406 if (!slob_last(prev) && b + units == next) { slob_free()
407 units += slob_units(next); slob_free()
408 set_slob(b, units, slob_next(next)); slob_free()
410 set_slob(b, units, next); slob_free()
/linux-4.4.14/include/linux/
H A Dtimerqueue.h15 struct timerqueue_node *next; member in struct:timerqueue_head
37 return head->next; timerqueue_getnext()
48 head->next = NULL; timerqueue_init_head()
H A Duser-return-notifier.h19 struct task_struct *next) propagate_user_return_notify()
23 set_tsk_thread_flag(next, TIF_USER_RETURN_NOTIFY); propagate_user_return_notify()
39 struct task_struct *next) propagate_user_return_notify()
18 propagate_user_return_notify(struct task_struct *prev, struct task_struct *next) propagate_user_return_notify() argument
38 propagate_user_return_notify(struct task_struct *prev, struct task_struct *next) propagate_user_return_notify() argument
H A Dlist_bl.h38 struct hlist_bl_node *next, **pprev; member in struct:hlist_bl_node
45 h->next = NULL; INIT_HLIST_BL_NODE()
81 n->next = first; hlist_bl_add_head()
83 first->pprev = &n->next; hlist_bl_add_head()
90 struct hlist_bl_node *next = n->next; __hlist_bl_del() local
98 ((unsigned long)next | __hlist_bl_del()
100 if (next) __hlist_bl_del()
101 next->pprev = pprev; __hlist_bl_del()
107 n->next = LIST_POISON1; hlist_bl_del()
146 pos = pos->next)
158 pos && ({ n = pos->next; 1; }) && \
H A Dlist_nulls.h25 struct hlist_nulls_node *next, **pprev; member in struct:hlist_nulls_node
68 n->next = first; hlist_nulls_add_head()
72 first->pprev = &n->next; hlist_nulls_add_head()
77 struct hlist_nulls_node *next = n->next; __hlist_nulls_del() local
80 WRITE_ONCE(*pprev, next); __hlist_nulls_del()
81 if (!is_a_nulls(next)) __hlist_nulls_del()
82 next->pprev = pprev; __hlist_nulls_del()
103 pos = pos->next)
115 pos = pos->next)
H A Dlist.h15 * sometimes we already know the next/prev entries and we can
27 list->next = list; INIT_LIST_HEAD()
35 * the prev/next entries already!
40 struct list_head *next) __list_add()
42 next->prev = new; __list_add()
43 new->next = next; __list_add()
45 prev->next = new; __list_add()
50 struct list_head *next);
63 __list_add(new, head, head->next); list_add()
81 * Delete a list entry by making the prev/next entries
85 * the prev/next entries already!
87 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument
89 next->prev = prev; __list_del()
90 WRITE_ONCE(prev->next, next); __list_del()
102 __list_del(entry->prev, entry->next); __list_del_entry()
107 __list_del(entry->prev, entry->next); list_del()
108 entry->next = LIST_POISON1; list_del()
126 new->next = old->next; list_replace()
127 new->next->prev = new; list_replace()
129 new->prev->next = new; list_replace()
180 return list->next == head; list_is_last()
189 return head->next == head; list_empty()
198 * in the process of modifying either member (next or prev)
207 struct list_head *next = head->next; list_empty_careful() local
208 return (next == head) && (next == head->prev); list_empty_careful()
220 first = head->next; list_rotate_left()
231 return !list_empty(head) && (head->next == head->prev); list_is_singular()
237 struct list_head *new_first = entry->next; __list_cut_position()
238 list->next = head->next; __list_cut_position()
239 list->next->prev = list; __list_cut_position()
241 entry->next = list; __list_cut_position()
242 head->next = new_first; __list_cut_position()
266 (head->next != entry && head != entry)) list_cut_position()
276 struct list_head *next) __list_splice()
278 struct list_head *first = list->next; __list_splice()
282 prev->next = first; __list_splice()
284 last->next = next; __list_splice()
285 next->prev = last; __list_splice()
297 __list_splice(list, head, head->next); list_splice()
323 __list_splice(list, head, head->next); list_splice_init()
363 list_entry((ptr)->next, type, member)
388 * list_next_entry - get the next element in list
393 list_entry((pos)->member.next, typeof(*(pos)), member)
409 for (pos = (head)->next; pos != (head); pos = pos->next)
426 for (pos = (head)->next, n = pos->next; pos != (head); \
427 pos = n, n = pos->next)
600 h->next = NULL; INIT_HLIST_NODE()
616 struct hlist_node *next = n->next; __hlist_del() local
619 WRITE_ONCE(*pprev, next); __hlist_del()
620 if (next) __hlist_del()
621 next->pprev = pprev; __hlist_del()
627 n->next = LIST_POISON1; hlist_del()
642 n->next = first; hlist_add_head()
644 first->pprev = &n->next; hlist_add_head()
649 /* next must be != NULL */ hlist_add_before()
651 struct hlist_node *next) hlist_add_before()
653 n->pprev = next->pprev; hlist_add_before()
654 n->next = next; hlist_add_before()
655 next->pprev = &n->next; hlist_add_before()
662 n->next = prev->next; hlist_add_behind()
663 prev->next = n; hlist_add_behind()
664 n->pprev = &prev->next; hlist_add_behind()
666 if (n->next) hlist_add_behind()
667 n->next->pprev = &n->next; hlist_add_behind()
673 n->pprev = &n->next; hlist_add_fake()
678 return h->pprev == &h->next; hlist_fake()
697 for (pos = (head)->first; pos ; pos = pos->next)
700 for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \
717 pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
725 for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\
727 pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
736 pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member))
747 pos && ({ n = pos->member.next; 1; }); \
38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
650 hlist_add_before(struct hlist_node *n, struct hlist_node *next) hlist_add_before() argument
H A Drculist.h32 WRITE_ONCE(list->next, list); INIT_LIST_HEAD_RCU()
37 * return the ->next pointer of a list_head in an rcu safe
40 #define list_next_rcu(list) (*((struct list_head __rcu **)(&(list)->next)))
46 * the prev/next entries already!
50 struct list_head *prev, struct list_head *next) __list_add_rcu()
52 new->next = next; __list_add_rcu()
55 next->prev = new; __list_add_rcu()
59 struct list_head *prev, struct list_head *next);
80 __list_add_rcu(new, head, head->next); list_add_rcu()
174 new->next = old->next; list_replace_rcu()
177 new->next->prev = new; list_replace_rcu()
202 struct list_head *first = list->next; list_splice_init_rcu()
204 struct list_head *at = head->next; list_splice_init_rcu()
234 last->next = at; list_splice_init_rcu()
265 * list_first_entry_rcu rereads the ->next pointer.
267 * Rereading the ->next pointer is not a problem for list_empty() and
288 struct list_head *__next = READ_ONCE(__ptr->next); \
303 for (pos = list_entry_rcu((head)->next, typeof(*pos), member); \
305 pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
317 for (pos = list_entry_rcu(pos->member.next, typeof(*pos), member); \
319 pos = list_entry_rcu(pos->member.next, typeof(*pos), member))
356 struct hlist_node *next = old->next; hlist_replace_rcu() local
358 new->next = next; hlist_replace_rcu()
361 if (next) hlist_replace_rcu()
362 new->next->pprev = &new->next; hlist_replace_rcu()
367 * return the first or the next element in an RCU protected hlist
370 #define hlist_next_rcu(node) (*((struct hlist_node __rcu **)(&(node)->next)))
397 n->next = first; hlist_add_head_rcu()
401 first->pprev = &n->next; hlist_add_head_rcu()
407 * @next: the existing element to add the new element before.
423 struct hlist_node *next) hlist_add_before_rcu()
425 n->pprev = next->pprev; hlist_add_before_rcu()
426 n->next = next; hlist_add_before_rcu()
428 next->pprev = &n->next; hlist_add_before_rcu()
452 n->next = prev->next; hlist_add_behind_rcu()
453 n->pprev = &prev->next; hlist_add_behind_rcu()
455 if (n->next) hlist_add_behind_rcu()
456 n->next->pprev = &n->next; hlist_add_behind_rcu()
49 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
422 hlist_add_before_rcu(struct hlist_node *n, struct hlist_node *next) hlist_add_before_rcu() argument
H A Dif_tunnel.h14 for (pos = rcu_dereference(start); pos; pos = rcu_dereference(pos->next))
H A Dihex.h23 /* Find the next record, taking into account the 4-byte alignment */
27 int next = ((be16_to_cpu(rec->len) + 5) & ~3) - 2; ihex_next_binrec() local
28 rec = (void *)&rec->data[next]; ihex_next_binrec()
46 /* Point to next record... */ ihex_validate_fw()
H A Dinitrd.h16 /* free_initrd_mem always gets called with the next two as arguments.. */
/linux-4.4.14/lib/lzo/
H A Dlzo1x_decompress_safe.c43 size_t t, next; lzo1x_decompress_safe() local
57 next = t; lzo1x_decompress_safe()
110 next = t & 3; lzo1x_decompress_safe()
121 next = t & 3; lzo1x_decompress_safe()
128 next = t & 3; lzo1x_decompress_safe()
152 next = get_unaligned_le16(ip); lzo1x_decompress_safe()
154 m_pos -= next >> 2; lzo1x_decompress_safe()
155 next &= 3; lzo1x_decompress_safe()
176 next = get_unaligned_le16(ip); lzo1x_decompress_safe()
178 m_pos -= next >> 2; lzo1x_decompress_safe()
179 next &= 3; lzo1x_decompress_safe()
199 state = next; lzo1x_decompress_safe()
201 op += next; lzo1x_decompress_safe()
202 ip += next; lzo1x_decompress_safe()
225 state = next; lzo1x_decompress_safe()
226 t = next; lzo1x_decompress_safe()
/linux-4.4.14/arch/sh/mm/
H A Dtlb-sh5.c27 cpu_data->dtlb.next = cpu_data->dtlb.first; sh64_tlb_init()
38 cpu_data->itlb.next = cpu_data->itlb.first; sh64_tlb_init()
47 * sh64_next_free_dtlb_entry - Find the next available DTLB entry
51 return cpu_data->dtlb.next; sh64_next_free_dtlb_entry()
62 cpu_data->dtlb.next += cpu_data->dtlb.step; sh64_get_wired_dtlb_entry()
104 cpu_data->dtlb.next = entry; sh64_put_wired_dtlb_entry()
191 unsigned long long next; __update_tlb() local
214 next = tlbp->next; __update_tlb()
215 __flush_tlb_slot(next); __update_tlb()
218 : : "r" (next), "r" (pteh), "r" (ptel) ); __update_tlb()
220 next += TLB_STEP; __update_tlb()
221 if (next > tlbp->last) __update_tlb()
222 next = tlbp->first; __update_tlb()
223 tlbp->next = next; __update_tlb()
H A Dgup.c122 unsigned long next; gup_pmd_range() local
129 next = pmd_addr_end(addr, end); gup_pmd_range()
132 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) gup_pmd_range()
134 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
142 unsigned long next; gup_pud_range() local
149 next = pud_addr_end(addr, end); gup_pud_range()
152 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) gup_pud_range()
154 } while (pudp++, addr = next, addr != end); gup_pud_range()
168 unsigned long next; __get_user_pages_fast() local
190 next = pgd_addr_end(addr, end); __get_user_pages_fast()
193 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) __get_user_pages_fast()
195 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
222 unsigned long next; get_user_pages_fast() local
239 next = pgd_addr_end(addr, end); get_user_pages_fast()
242 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) get_user_pages_fast()
244 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
/linux-4.4.14/tools/include/linux/
H A Dlist.h19 begin->prev->next = end->next; list_del_range()
20 end->next->prev = begin->prev; list_del_range()
29 for (; pos != (head); pos = pos->next)
/linux-4.4.14/arch/x86/include/asm/
H A Dswitch_to.h6 struct task_struct *next);
15 "movl %P[task_canary](%[next]), %%ebx\n\t" \
31 #define switch_to(prev, next, last) \
66 : [next_sp] "m" (next->thread.sp), \
67 [next_ip] "m" (next->thread.ip), \
71 [next] "d" (next) \
108 #define switch_to(prev, next, last) \
111 "movq %P[threadrsp](%[next]),%%rsp\n\t" /* restore RSP */ \
122 : [next] "S" (next), [prev] "D" (prev), \
H A Dmmu_context.h16 struct mm_struct *next) paravirt_activate_mm()
107 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
112 if (likely(prev != next)) { switch_mm()
115 this_cpu_write(cpu_tlbstate.active_mm, next); switch_mm()
117 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm()
124 * CPU 0: Write to a PTE for 'next' switch_mm()
126 * CPU 1: set bit 1 in next's mm_cpumask switch_mm()
140 * from next->pgd. TLB fills are special and can happen switch_mm()
147 load_cr3(next->pgd); switch_mm()
155 load_mm_cr4(next); switch_mm()
167 * exists. That means that next->context.ldt != switch_mm()
170 if (unlikely(prev->context.ldt != next->context.ldt)) switch_mm()
171 load_mm_ldt(next); switch_mm()
177 BUG_ON(this_cpu_read(cpu_tlbstate.active_mm) != next); switch_mm()
179 if (!cpumask_test_cpu(cpu, mm_cpumask(next))) { switch_mm()
186 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm()
196 load_cr3(next->pgd); switch_mm()
198 load_mm_cr4(next); switch_mm()
199 load_mm_ldt(next); switch_mm()
205 #define activate_mm(prev, next) \
207 paravirt_activate_mm((prev), (next)); \
208 switch_mm((prev), (next), NULL); \
15 paravirt_activate_mm(struct mm_struct *prev, struct mm_struct *next) paravirt_activate_mm() argument
/linux-4.4.14/arch/arm/mm/
H A Didmap.c25 unsigned long next; idmap_add_pmd() local
46 next = pmd_addr_end(addr, end); idmap_add_pmd()
49 } while (pmd++, addr = next, addr != end); idmap_add_pmd()
69 unsigned long next; idmap_add_pud() local
72 next = pud_addr_end(addr, end); idmap_add_pud()
73 idmap_add_pmd(pud, addr, next, prot); idmap_add_pud()
74 } while (pud++, addr = next, addr != end); idmap_add_pud()
81 unsigned long next; identity_mapping_add() local
94 next = pgd_addr_end(addr, end); identity_mapping_add()
95 idmap_add_pud(pgd, addr, next, prot); identity_mapping_add()
96 } while (pgd++, addr = next, addr != end); identity_mapping_add()
/linux-4.4.14/arch/cris/arch-v32/mm/
H A Dintmem.c110 struct intmem_allocation *next = crisv32_intmem_free() local
111 list_entry(allocation->entry.next, crisv32_intmem_free()
115 /* Join with prev and/or next if also free */ crisv32_intmem_free()
123 if ((next != &intmem_allocations) && crisv32_intmem_free()
124 (next->status == STATUS_FREE)) { crisv32_intmem_free()
125 allocation->size += next->size; crisv32_intmem_free()
126 list_del(&next->entry); crisv32_intmem_free()
127 kfree(next); crisv32_intmem_free()
/linux-4.4.14/arch/metag/include/asm/
H A Dmmu_context.h72 static inline void switch_mmu(struct mm_struct *prev, struct mm_struct *next) switch_mmu() argument
75 load_pgd(next->pgd, hard_processor_id()); switch_mmu()
87 next->pgd = prev->pgd; switch_mmu()
91 next->pgd[i] = ((pgd_t *) next->context.pgd_base)[i]; switch_mmu()
98 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
101 if (prev != next) switch_mm()
102 switch_mmu(prev, next); switch_mm()
/linux-4.4.14/drivers/usb/host/
H A Dohci-dbg.c22 #define ohci_dbg_sw(ohci, next, size, format, arg...) \
24 if (next != NULL) { \
26 s_len = scnprintf (*next, *size, format, ## arg ); \
27 *size -= s_len; *next += s_len; \
32 /* Version for use where "next" is the address of a local variable */
33 #define ohci_dbg_nosw(ohci, next, size, format, arg...) \
36 s_len = scnprintf(*next, *size, format, ## arg); \
37 *size -= s_len; *next += s_len; \
45 char **next, ohci_dump_intr_mask()
48 ohci_dbg_sw (ohci, next, size, "%s 0x%08x%s%s%s%s%s%s%s%s%s\n", ohci_dump_intr_mask()
67 char **next, maybe_print_eds()
71 ohci_dbg_sw (ohci, next, size, "%s %08x\n", label, value); maybe_print_eds()
100 ohci_dump_status (struct ohci_hcd *controller, char **next, unsigned *size) ohci_dump_status() argument
106 ohci_dbg_sw (controller, next, size, ohci_dump_status()
113 ohci_dbg_sw (controller, next, size, ohci_dump_status()
128 ohci_dbg_sw (controller, next, size, ohci_dump_status()
139 next, size); ohci_dump_status()
142 next, size); ohci_dump_status()
147 next, size); ohci_dump_status()
151 next, size); ohci_dump_status()
154 next, size); ohci_dump_status()
158 next, size); ohci_dump_status()
161 next, size); ohci_dump_status()
164 ohci_readl (controller, &regs->donehead), next, size); ohci_dump_status()
167 #define dbg_port_sw(hc,num,value,next,size) \
168 ohci_dbg_sw (hc, next, size, \
193 char **next, ohci_dump_roothub()
203 ohci_dbg_sw (controller, next, size, ohci_dump_roothub()
214 ohci_dbg_sw (controller, next, size, ohci_dump_roothub()
221 ohci_dbg_sw (controller, next, size, ohci_dump_roothub()
235 dbg_port_sw (controller, i, temp, next, size); ohci_dump_roothub()
259 ohci_dbg (ohci, "%s td %p%s; urb %p index %d; hw next td %08x\n", ohci_dump_td()
317 ohci_dbg (ohci, "%s, ed %p state 0x%x type %s; next ed %08x\n", ohci_dump_ed()
491 char *next; fill_periodic_buffer() local
500 next = buf->page; fill_periodic_buffer()
503 temp = scnprintf (next, size, "size = %d\n", NUM_INTS); fill_periodic_buffer()
505 next += temp; fill_periodic_buffer()
514 temp = scnprintf (next, size, "%2d [%3d]:", i, ohci->load [i]); fill_periodic_buffer()
516 next += temp; fill_periodic_buffer()
519 temp = scnprintf (next, size, " ed%d/%p", fill_periodic_buffer()
522 next += temp; fill_periodic_buffer()
538 temp = scnprintf (next, size, fill_periodic_buffer()
554 next += temp; fill_periodic_buffer()
569 temp = scnprintf (next, size, "\n"); fill_periodic_buffer()
571 next += temp; fill_periodic_buffer()
587 char *next; fill_registers_buffer() local
593 next = buf->page; fill_registers_buffer()
600 ohci_dbg_nosw(ohci, &next, &size, fill_registers_buffer()
610 size -= scnprintf (next, size, fill_registers_buffer()
615 ohci_dump_status(ohci, &next, &size); fill_registers_buffer()
619 ohci_dbg_nosw(ohci, &next, &size, fill_registers_buffer()
624 temp = scnprintf (next, size, fill_registers_buffer()
629 next += temp; fill_registers_buffer()
632 temp = scnprintf (next, size, "fmremaining 0x%08x %sFR=0x%04x\n", fill_registers_buffer()
636 next += temp; fill_registers_buffer()
639 temp = scnprintf (next, size, "periodicstart 0x%04x\n", fill_registers_buffer()
642 next += temp; fill_registers_buffer()
645 temp = scnprintf (next, size, "lsthresh 0x%04x\n", fill_registers_buffer()
648 next += temp; fill_registers_buffer()
650 temp = scnprintf (next, size, "hub poll timer %s\n", fill_registers_buffer()
653 next += temp; fill_registers_buffer()
656 ohci_dump_roothub (ohci, 1, &next, &size); fill_registers_buffer()
41 ohci_dump_intr_mask( struct ohci_hcd *ohci, char *label, u32 mask, char **next, unsigned *size) ohci_dump_intr_mask() argument
63 maybe_print_eds( struct ohci_hcd *ohci, char *label, u32 value, char **next, unsigned *size) maybe_print_eds() argument
190 ohci_dump_roothub( struct ohci_hcd *controller, int verbose, char **next, unsigned *size) ohci_dump_roothub() argument
H A Dehci-dbg.c135 ehci_dbg (ehci, "%s [%d] itd %p, next %08x, urb %p\n", dbg_itd()
166 ehci_dbg (ehci, "%s [%d] sitd %p, next %08x, urb %p\n", dbg_sitd()
419 char *next = *nextp; qh_lines() local
438 temp = scnprintf (next, size, qh_lines()
449 next += temp; qh_lines()
466 temp = snprintf (next, size, qh_lines()
481 next += temp; qh_lines()
486 temp = snprintf (next, size, "\n"); qh_lines()
490 next += temp; qh_lines()
494 *nextp = next; qh_lines()
503 char *next; fill_async_buffer() local
508 next = buf->output_buf; fill_async_buffer()
511 *next = 0; fill_async_buffer()
519 qh_lines (ehci, qh, &next, &size); fill_async_buffer()
521 temp = scnprintf(next, size, "\nunlink =\n"); fill_async_buffer()
523 next += temp; fill_async_buffer()
528 qh_lines(ehci, qh, &next, &size); fill_async_buffer()
542 char *next; fill_bandwidth_buffer() local
549 next = buf->output_buf; fill_bandwidth_buffer()
552 *next = 0; fill_bandwidth_buffer()
557 temp = scnprintf(next, size, fill_bandwidth_buffer()
560 next += temp; fill_bandwidth_buffer()
563 temp = scnprintf(next, size, fill_bandwidth_buffer()
568 next += temp; fill_bandwidth_buffer()
573 temp = scnprintf(next, size, fill_bandwidth_buffer()
578 next += temp; fill_bandwidth_buffer()
581 temp = scnprintf(next, size, fill_bandwidth_buffer()
586 next += temp; fill_bandwidth_buffer()
588 temp = scnprintf(next, size, fill_bandwidth_buffer()
591 next += temp; fill_bandwidth_buffer()
595 temp = scnprintf(next, size, fill_bandwidth_buffer()
600 next += temp; fill_bandwidth_buffer()
603 temp = scnprintf(next, size, fill_bandwidth_buffer()
611 next += temp; fill_bandwidth_buffer()
616 return next - buf->output_buf; fill_bandwidth_buffer()
627 char *next; fill_periodic_buffer() local
638 next = buf->output_buf; fill_periodic_buffer()
641 temp = scnprintf (next, size, "size = %d\n", ehci->periodic_size); fill_periodic_buffer()
643 next += temp; fill_periodic_buffer()
655 temp = scnprintf (next, size, "%4d: ", i); fill_periodic_buffer()
657 next += temp; fill_periodic_buffer()
665 temp = scnprintf (next, size, " qh%d-%04x/%p", fill_periodic_buffer()
673 next += temp; fill_periodic_buffer()
679 temp = scnprintf (next, size, fill_periodic_buffer()
682 next += temp; fill_periodic_buffer()
707 temp = scnprintf (next, size, fill_periodic_buffer()
726 temp = scnprintf (next, size, fill_periodic_buffer()
733 temp = scnprintf (next, size, fill_periodic_buffer()
739 temp = scnprintf (next, size, fill_periodic_buffer()
750 next += temp; fill_periodic_buffer()
753 temp = scnprintf (next, size, "\n"); fill_periodic_buffer()
755 next += temp; fill_periodic_buffer()
785 char *next, scratch [80]; fill_registers_buffer() local
791 next = buf->output_buf; fill_registers_buffer()
797 size = scnprintf (next, size, fill_registers_buffer()
809 temp = scnprintf (next, size, fill_registers_buffer()
818 next += temp; fill_registers_buffer()
834 temp = scnprintf (next, size, fill_registers_buffer()
839 next += temp; fill_registers_buffer()
843 temp = scnprintf (next, size, fill_registers_buffer()
846 next += temp; fill_registers_buffer()
861 temp = scnprintf (next, size, "structural params 0x%08x\n", i); fill_registers_buffer()
863 next += temp; fill_registers_buffer()
866 temp = scnprintf (next, size, "capability params 0x%08x\n", i); fill_registers_buffer()
868 next += temp; fill_registers_buffer()
873 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer()
875 next += temp; fill_registers_buffer()
879 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer()
881 next += temp; fill_registers_buffer()
885 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer()
887 next += temp; fill_registers_buffer()
889 temp = scnprintf (next, size, "uframe %04x\n", fill_registers_buffer()
892 next += temp; fill_registers_buffer()
898 temp = scnprintf (next, size, fmt, temp, scratch); fill_registers_buffer()
900 next += temp; fill_registers_buffer()
902 temp = scnprintf (next, size, fill_registers_buffer()
907 next += temp; fill_registers_buffer()
912 temp = scnprintf(next, size, "async unlink qh %p\n", fill_registers_buffer()
916 next += temp; fill_registers_buffer()
920 temp = scnprintf (next, size, fill_registers_buffer()
925 next += temp; fill_registers_buffer()
927 temp = scnprintf (next, size, "complete %ld unlink %ld\n", fill_registers_buffer()
930 next += temp; fill_registers_buffer()
H A Dxhci-ext-caps.h94 * Return the next extended capability pointer register.
100 * in XHCI_HCC_PARAMS_OFFSET. If searching for the next extended capability,
103 * Returns 0 if there is no next extended capability register or returns the register offset
108 u32 next; xhci_find_next_cap_offset() local
110 next = readl(base + ext_offset); xhci_find_next_cap_offset()
114 next = XHCI_HCC_EXT_CAPS(next); xhci_find_next_cap_offset()
117 /* Find the next extended capability */ xhci_find_next_cap_offset()
118 next = XHCI_EXT_CAPS_NEXT(next); xhci_find_next_cap_offset()
121 if (!next) xhci_find_next_cap_offset()
127 return ext_offset + (next << 2); xhci_find_next_cap_offset()
H A Dfhci-q.c79 td = list_entry(ed->td_list.next, struct td, node); peek_td_from_ed()
91 td = list_entry(frame->tds_list.next, struct td, frame_lh); fhci_remove_td_from_frame()
92 list_del_init(frame->tds_list.next); fhci_remove_td_from_frame()
104 td = list_entry(frame->tds_list.next, struct td, frame_lh); fhci_peek_td_from_frame()
116 td = list_entry(ed->td_list.next, struct td, node); fhci_remove_td_from_ed()
117 list_del_init(ed->td_list.next); fhci_remove_td_from_ed()
119 /* if this TD was the ED's head, find next TD */ fhci_remove_td_from_ed()
121 ed->td_head = list_entry(ed->td_list.next, struct td, fhci_remove_td_from_ed()
136 td = list_entry(p_list->done_list.next, struct td, node); fhci_remove_td_from_done_list()
137 list_del_init(p_list->done_list.next); fhci_remove_td_from_done_list()
151 /* If this TD was the ED's head,find next TD */ fhci_move_td_from_ed_to_done_list()
153 ed->td_head = list_entry(ed->td_list.next, struct td, node); fhci_move_td_from_ed_to_done_list()
176 /* if this TD was the ED's head,find the next TD */ free_urb_priv()
178 ed->td_head = list_entry(ed->td_list.next, struct td, node); free_urb_priv()
186 /* if this TD was the ED's head,find next TD */ free_urb_priv()
/linux-4.4.14/scripts/gdb/linux/
H A Dlists.py35 n = c['next'].dereference()
37 if p['next'] != c.address:
38 gdb.write('prev.next != current: '
56 gdb.write('next.prev != current: '
58 'next@{n_addr}={n}\n'.format(
66 gdb.write('next is not accessible: '
H A Dmodules.py26 entry = modules['next']
31 entry = entry['next']
84 entry = source_list['next']
92 entry = entry['next']
/linux-4.4.14/arch/mn10300/include/asm/
H A Dswitch_to.h22 #define switch_fpu(prev, next) \
31 #define switch_fpu(prev, next) do {} while (0)
37 struct thread_struct *next,
40 #define switch_to(prev, next, last) \
42 switch_fpu(prev, next); \
44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
H A Dmmu_context.h140 * @next: The incoming MM context.
143 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
148 if (prev != next) { switch_mm()
150 per_cpu(cpu_tlbstate, cpu).active_mm = next; switch_mm()
152 cpu_ran_vm(cpu, next); switch_mm()
153 PTBR = (unsigned long) next->pgd; switch_mm()
154 activate_context(next); switch_mm()
159 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
/linux-4.4.14/sound/core/seq/
H A Dseq_prioq.c38 * |next
168 f->tail->next = cell; snd_seq_prioq_cell_in()
170 cell->next = NULL; snd_seq_prioq_cell_in()
193 /* move cursor to next cell */ snd_seq_prioq_cell_in()
195 cur = cur->next; snd_seq_prioq_cell_in()
205 prev->next = cell; snd_seq_prioq_cell_in()
206 cell->next = cur; snd_seq_prioq_cell_in()
231 f->head = cell->next; snd_seq_prioq_cell_out()
237 cell->next = NULL; snd_seq_prioq_cell_out()
292 register struct snd_seq_event_cell *cell, *next; snd_seq_prioq_leave() local
301 next = cell->next; snd_seq_prioq_leave()
305 f->head = cell->next; snd_seq_prioq_leave()
307 prev->next = cell->next; snd_seq_prioq_leave()
310 f->tail = cell->next; snd_seq_prioq_leave()
313 cell->next = NULL; snd_seq_prioq_leave()
317 freeprev->next = cell; snd_seq_prioq_leave()
331 cell = next; snd_seq_prioq_leave()
337 freenext = freefirst->next; snd_seq_prioq_leave()
402 struct snd_seq_event_cell *cell, *next; snd_seq_prioq_remove_events() local
412 next = cell->next; snd_seq_prioq_remove_events()
418 f->head = cell->next; snd_seq_prioq_remove_events()
420 prev->next = cell->next; snd_seq_prioq_remove_events()
424 f->tail = cell->next; snd_seq_prioq_remove_events()
428 cell->next = NULL; snd_seq_prioq_remove_events()
432 freeprev->next = cell; snd_seq_prioq_remove_events()
439 cell = next; snd_seq_prioq_remove_events()
445 freenext = freefirst->next; snd_seq_prioq_remove_events()
H A Dseq_fifo.c134 f->tail->next = cell; snd_seq_fifo_event_in()
157 f->head = cell->next; fifo_cell_out()
163 cell->next = NULL; fifo_cell_out()
215 cell->next = f->head; snd_seq_fifo_cell_putback()
236 struct snd_seq_event_cell *cell, *next, *oldhead; snd_seq_fifo_resize() local
263 for (cell = oldhead; cell; cell = next) { snd_seq_fifo_resize()
264 next = cell->next; snd_seq_fifo_resize()
/linux-4.4.14/arch/mn10300/kernel/
H A Dkprobes.c125 * Calculate what the PC will be after executing next instruction
133 u8 opc, *pc, *sp, *next; find_nextpc() local
135 next = 0; find_nextpc()
144 next = pc + size; find_nextpc()
152 next = pc + x8; find_nextpc()
160 next = pc + x16; find_nextpc()
168 next = pc + x32; find_nextpc()
174 next = (u8 *)regs->mdr; find_nextpc()
183 next = (u8 *)x32; find_nextpc()
189 next = pc + 2; find_nextpc()
195 next = (u8 *)regs->a0; find_nextpc()
198 next = (u8 *)regs->a1; find_nextpc()
201 next = (u8 *)regs->a2; find_nextpc()
204 next = (u8 *)regs->a3; find_nextpc()
212 next = (u8 *)x32; find_nextpc()
218 next = (u8 *)x32; find_nextpc()
226 next = pc + 3; find_nextpc()
233 next = pc + x8; find_nextpc()
242 next = pc + x16; find_nextpc()
244 next = pc + 4; find_nextpc()
254 next = pc + x32; find_nextpc()
258 /* SETLB - loads the next four bytes into the LIR reg */ find_nextpc()
265 return (unsigned)next; find_nextpc()
274 u8 opc, *pc, *sp, *next; singlestep_branch_setup() local
276 next = NULL; singlestep_branch_setup()
286 next = pc; singlestep_branch_setup()
292 next = pc + 5; singlestep_branch_setup()
300 next = pc + 7; singlestep_branch_setup()
304 next = pc + 3; singlestep_branch_setup()
305 regs->mdr = (unsigned) next; singlestep_branch_setup()
310 next = pc + 3; singlestep_branch_setup()
311 *(unsigned *)sp = (unsigned) next; singlestep_branch_setup()
315 next = pc + 2; singlestep_branch_setup()
324 next = pc + 4; singlestep_branch_setup()
327 next = pc; singlestep_branch_setup()
330 next = pc + 2; singlestep_branch_setup()
331 *(unsigned *) sp = (unsigned) next; singlestep_branch_setup()
334 next = pc + 2; singlestep_branch_setup()
335 *(unsigned *)(sp + 4) = (unsigned) next; singlestep_branch_setup()
342 next = pc + 4; singlestep_branch_setup()
350 next = pc + 6; singlestep_branch_setup()
358 return (unsigned) next; singlestep_branch_setup()
/linux-4.4.14/arch/arm64/mm/
H A Dkasan_init.c30 unsigned long next; kasan_early_pte_populate() local
37 next = addr + PAGE_SIZE; kasan_early_pte_populate()
40 } while (pte++, addr = next, addr != end && pte_none(*pte)); kasan_early_pte_populate()
48 unsigned long next; kasan_early_pmd_populate() local
55 next = pmd_addr_end(addr, end); kasan_early_pmd_populate()
56 kasan_early_pte_populate(pmd, addr, next); kasan_early_pmd_populate()
57 } while (pmd++, addr = next, addr != end && pmd_none(*pmd)); kasan_early_pmd_populate()
65 unsigned long next; kasan_early_pud_populate() local
72 next = pud_addr_end(addr, end); kasan_early_pud_populate()
73 kasan_early_pmd_populate(pud, addr, next); kasan_early_pud_populate()
74 } while (pud++, addr = next, addr != end && pud_none(*pud)); kasan_early_pud_populate()
81 unsigned long next; kasan_map_early_shadow() local
86 next = pgd_addr_end(addr, end); kasan_map_early_shadow()
87 kasan_early_pud_populate(pgd, addr, next); kasan_map_early_shadow()
88 } while (pgd++, addr = next, addr != end); kasan_map_early_shadow()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/core/
H A Dmm.c54 struct nvkm_mm_node *next = node(this, next); nvkm_mm_free() local
62 if (next && next->type == NVKM_MM_TYPE_NONE) { nvkm_mm_free()
63 next->offset = this->offset; nvkm_mm_free()
64 next->length += this->length; nvkm_mm_free()
114 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_head() local
133 next = node(this, next); nvkm_mm_head()
134 if (next && next->type != type) nvkm_mm_head()
188 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_tail() local
206 next = node(this, next); nvkm_mm_tail()
207 if (next && next->type != type) { nvkm_mm_tail()
209 c = next->offset - e; nvkm_mm_tail()
241 u32 next; nvkm_mm_init() local
245 next = prev->offset + prev->length; nvkm_mm_init()
246 if (next != offset) { nvkm_mm_init()
247 BUG_ON(next > offset); nvkm_mm_init()
251 node->offset = next; nvkm_mm_init()
252 node->length = offset - next; nvkm_mm_init()
/linux-4.4.14/arch/nios2/mm/
H A Dmmu_context.c60 /* Return the next pid */ get_new_context()
80 void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
89 if (unlikely(CTX_VERSION(next->context) != switch_mm()
91 next->context = get_new_context(); switch_mm()
94 pgd_current = next->pgd; switch_mm()
97 set_context(next->context); switch_mm()
106 void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
108 next->context = get_new_context(); activate_mm()
109 set_context(next->context); activate_mm()
110 pgd_current = next->pgd; activate_mm()
/linux-4.4.14/arch/powerpc/include/asm/
H A Dmmu_context.h37 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next);
51 extern void switch_cop(struct mm_struct *next); mmu_context_init()
59 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
63 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); switch_mm()
67 tsk->thread.pgdir = next->pgd; switch_mm()
72 get_paca()->pgd = next->pgd; switch_mm()
75 if (prev == next) switch_mm()
79 /* Switch coprocessor context only if prev or next uses a coprocessor */ switch_mm()
80 if (prev->context.acop || next->context.acop) switch_mm()
81 switch_cop(next); switch_mm()
96 switch_slb(tsk, next); switch_mm()
99 switch_mmu_context(prev, next); switch_mm()
110 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
115 switch_mm(prev, next, current); activate_mm()
H A Dbootx.h28 u32 next; member in struct:bootx_dt_prop
44 u32 next; member in struct:bootx_dt_node
/linux-4.4.14/drivers/parisc/
H A Dhppb.c29 struct hppb_card *next; member in struct:hppb_card
34 .next = NULL,
53 while(card->next) { hppb_probe()
54 card = card->next; hppb_probe()
58 card->next = kzalloc(sizeof(struct hppb_card), GFP_KERNEL); hppb_probe()
59 if(!card->next) { hppb_probe()
63 card = card->next; hppb_probe()
/linux-4.4.14/drivers/gpu/drm/nouveau/include/nvif/
H A Dlist.h12 * The above copyright notice and this permission notice (including the next
92 * struct foo *iterator, *next;
93 * list_for_each_entry_safe(iterator, next, &bar.list_of_foos, entry) {
111 struct list_head *next, *prev; member in struct:list_head
130 list->next = list->prev = list; INIT_LIST_HEAD()
135 struct list_head *prev, struct list_head *next) __list_add()
137 next->prev = entry; __list_add()
138 entry->next = next; __list_add()
140 prev->next = entry; __list_add()
161 __list_add(entry, head, head->next); list_add()
186 __list_del(struct list_head *prev, struct list_head *next) __list_del() argument
188 next->prev = prev; __list_del()
189 prev->next = next; __list_del()
209 __list_del(entry->prev, entry->next); list_del()
215 __list_del(entry->prev, entry->next); list_del_init()
222 __list_del(list->prev, list->next); list_move_tail()
237 return head->next == head; list_empty()
277 list_entry((ptr)->next, type, member)
315 for (pos = __container_of((head)->next, pos, member); \
317 pos = __container_of(pos->member.next, pos, member))
327 for (pos = __container_of((head)->next, pos, member), \
328 tmp = __container_of(pos->member.next, pos, member); \
330 pos = tmp, tmp = __container_of(pos->member.next, tmp, member))
339 for (pos = __container_of(pos->member.next, pos, member); \
341 pos = __container_of(pos->member.next, pos, member))
351 pos = __container_of(pos->member.next, pos, member))
134 __list_add(struct list_head *entry, struct list_head *prev, struct list_head *next) __list_add() argument
/linux-4.4.14/arch/tile/include/asm/
H A Dswitch_to.h36 #define prepare_arch_switch(next) _prepare_arch_switch(next)
37 void _prepare_arch_switch(struct task_struct *next);
40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next)))
42 struct task_struct *next);
46 struct task_struct *next,
H A Dcacheflush.h81 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); __flush_buffer() local
83 while (next < finish) { __flush_buffer()
84 __insn_flush(next); __flush_buffer()
85 next += CHIP_FLUSH_STRIDE(); __flush_buffer()
92 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); __finv_buffer() local
94 while (next < finish) { __finv_buffer()
95 __insn_finv(next); __finv_buffer()
96 next += CHIP_FINV_STRIDE(); __finv_buffer()
125 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); __inv_buffer() local
127 while (next < finish) { __inv_buffer()
128 __insn_inv(next); __inv_buffer()
129 next += CHIP_INV_STRIDE(); __inv_buffer()
/linux-4.4.14/arch/hexagon/include/asm/
H A Dmmu_context.h66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
75 if (next->context.generation < prev->context.generation) { switch_mm()
77 next->pgd[l1] = init_mm.pgd[l1]; switch_mm()
79 next->context.generation = prev->context.generation; switch_mm()
82 __vmnewmap((void *)next->context.ptbase); switch_mm()
88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
93 switch_mm(prev, next, current_thread_info()->task); activate_mm()
/linux-4.4.14/drivers/pci/hotplug/
H A Dcpqphp_ctrl.c135 slot = slot->next; cpqhp_find_slot()
321 if (!((*head)->next)) sort_by_size()
328 if (((*head)->next) && sort_by_size()
329 ((*head)->length > (*head)->next->length)) { sort_by_size()
332 *head = (*head)->next; sort_by_size()
333 current_res->next = (*head)->next; sort_by_size()
334 (*head)->next = current_res; sort_by_size()
339 while (current_res->next && current_res->next->next) { sort_by_size()
340 if (current_res->next->length > current_res->next->next->length) { sort_by_size()
342 next_res = current_res->next; sort_by_size()
343 current_res->next = current_res->next->next; sort_by_size()
344 current_res = current_res->next; sort_by_size()
345 next_res->next = current_res->next; sort_by_size()
346 current_res->next = next_res; sort_by_size()
348 current_res = current_res->next; sort_by_size()
369 if (!((*head)->next)) sort_by_max_size()
376 if (((*head)->next) && sort_by_max_size()
377 ((*head)->length < (*head)->next->length)) { sort_by_max_size()
380 *head = (*head)->next; sort_by_max_size()
381 current_res->next = (*head)->next; sort_by_max_size()
382 (*head)->next = current_res; sort_by_max_size()
387 while (current_res->next && current_res->next->next) { sort_by_max_size()
388 if (current_res->next->length < current_res->next->next->length) { sort_by_max_size()
390 next_res = current_res->next; sort_by_max_size()
391 current_res->next = current_res->next->next; sort_by_max_size()
392 current_res = current_res->next; sort_by_max_size()
393 next_res->next = current_res->next; sort_by_max_size()
394 current_res->next = next_res; sort_by_max_size()
396 current_res = current_res->next; sort_by_max_size()
460 split_node->next = node; do_pre_bridge_resource_split()
468 *head = node->next; do_pre_bridge_resource_split()
471 while (prevnode->next != node) do_pre_bridge_resource_split()
472 prevnode = prevnode->next; do_pre_bridge_resource_split()
474 prevnode->next = node->next; do_pre_bridge_resource_split()
476 node->next = NULL; do_pre_bridge_resource_split()
501 while (node->next) { do_bridge_resource_split()
503 node = node->next; do_bridge_resource_split()
556 for (node = *head; node; node = node->next) { get_io_resource()
581 split_node->next = node->next; get_io_resource()
582 node->next = split_node; get_io_resource()
600 split_node->next = node->next; get_io_resource()
601 node->next = split_node; get_io_resource()
612 *head = node->next; get_io_resource()
615 while (prevnode->next != node) get_io_resource()
616 prevnode = prevnode->next; get_io_resource()
618 prevnode->next = node->next; get_io_resource()
620 node->next = NULL; get_io_resource()
650 for (max = *head; max; max = max->next) { get_max_resource()
652 * instead we'll continue to the next. get_max_resource()
677 split_node->next = max->next; get_max_resource()
678 max->next = split_node; get_max_resource()
695 split_node->next = max->next; get_max_resource()
696 max->next = split_node; get_max_resource()
706 *head = max->next; get_max_resource()
708 while (temp && temp->next != max) get_max_resource()
709 temp = temp->next; get_max_resource()
712 temp->next = max->next; get_max_resource()
715 max->next = NULL; get_max_resource()
747 for (node = *head; node; node = node->next) { get_resource()
774 split_node->next = node->next; get_resource()
775 node->next = split_node; get_resource()
794 split_node->next = node->next; get_resource()
795 node->next = split_node; get_resource()
802 *head = node->next; get_resource()
805 while (prevnode->next != node) get_resource()
806 prevnode = prevnode->next; get_resource()
808 prevnode->next = node->next; get_resource()
810 node->next = NULL; get_resource()
838 dbg("*head->next = %p\n",(*head)->next); cpqhp_resource_sort_and_combine()
840 if (!(*head)->next) cpqhp_resource_sort_and_combine()
844 dbg("*head->next->base = 0x%x\n",(*head)->next->base); cpqhp_resource_sort_and_combine()
849 if (((*head)->next) && cpqhp_resource_sort_and_combine()
850 ((*head)->base > (*head)->next->base)) { cpqhp_resource_sort_and_combine()
852 (*head) = (*head)->next; cpqhp_resource_sort_and_combine()
853 node1->next = (*head)->next; cpqhp_resource_sort_and_combine()
854 (*head)->next = node1; cpqhp_resource_sort_and_combine()
860 while (node1->next && node1->next->next) { cpqhp_resource_sort_and_combine()
861 if (node1->next->base > node1->next->next->base) { cpqhp_resource_sort_and_combine()
863 node2 = node1->next; cpqhp_resource_sort_and_combine()
864 node1->next = node1->next->next; cpqhp_resource_sort_and_combine()
865 node1 = node1->next; cpqhp_resource_sort_and_combine()
866 node2->next = node1->next; cpqhp_resource_sort_and_combine()
867 node1->next = node2; cpqhp_resource_sort_and_combine()
869 node1 = node1->next; cpqhp_resource_sort_and_combine()
875 while (node1 && node1->next) { cpqhp_resource_sort_and_combine()
876 if ((node1->base + node1->length) == node1->next->base) { cpqhp_resource_sort_and_combine()
879 node1->length += node1->next->length; cpqhp_resource_sort_and_combine()
880 node2 = node1->next; cpqhp_resource_sort_and_combine()
881 node1->next = node1->next->next; cpqhp_resource_sort_and_combine()
884 node1 = node1->next; cpqhp_resource_sort_and_combine()
971 struct pci_func *next; cpqhp_slot_create() local
977 new_slot->next = NULL; cpqhp_slot_create()
983 next = cpqhp_slot_list[busnumber]; cpqhp_slot_create()
984 while (next->next != NULL) cpqhp_slot_create()
985 next = next->next; cpqhp_slot_create()
986 next->next = new_slot; cpqhp_slot_create()
1000 struct pci_func *next; slot_remove() local
1005 next = cpqhp_slot_list[old_slot->bus]; slot_remove()
1006 if (next == NULL) slot_remove()
1009 if (next == old_slot) { slot_remove()
1010 cpqhp_slot_list[old_slot->bus] = old_slot->next; slot_remove()
1016 while ((next->next != old_slot) && (next->next != NULL)) slot_remove()
1017 next = next->next; slot_remove()
1019 if (next->next == old_slot) { slot_remove()
1020 next->next = old_slot->next; slot_remove()
1039 struct pci_func *next; bridge_slot_remove() local
1045 next = cpqhp_slot_list[tempBus]; bridge_slot_remove()
1047 while (!slot_remove(next)) bridge_slot_remove()
1048 next = cpqhp_slot_list[tempBus]; bridge_slot_remove()
1051 next = cpqhp_slot_list[bridge->bus]; bridge_slot_remove()
1053 if (next == NULL) bridge_slot_remove()
1056 if (next == bridge) { bridge_slot_remove()
1057 cpqhp_slot_list[bridge->bus] = bridge->next; bridge_slot_remove()
1061 while ((next->next != bridge) && (next->next != NULL)) bridge_slot_remove()
1062 next = next->next; bridge_slot_remove()
1064 if (next->next != bridge) bridge_slot_remove()
1066 next->next = bridge->next; bridge_slot_remove()
1094 while (func->next != NULL) { cpqhp_slot_find()
1095 func = func->next; cpqhp_slot_find()
1144 for (slot = ctrl->slot; slot; slot = slot->next) { set_controller_speed()
1580 /* next, we will instantiate the linux pci_dev structures (with board_added()
1758 for (ctrl = cpqhp_ctrl_list; ctrl; ctrl=ctrl->next) event_thread()
2171 /* Get ready for next iteration */ switch_leds()
2222 /* Get ready for next iteration */ cpqhp_hardware_test()
2232 /* Get ready for next iteration */ cpqhp_hardware_test()
2316 /* The following loop skips to the next present function configure_new_device()
2459 dbg("(base, len, next) (%x, %x, %p)\n", io_node->base, configure_new_function()
2460 io_node->length, io_node->next); configure_new_function()
2462 dbg("(base, len, next) (%x, %x, %p)\n", mem_node->base, configure_new_function()
2463 mem_node->length, mem_node->next); configure_new_function()
2465 dbg("(base, len, next) (%x, %x, %p)\n", p_mem_node->base, configure_new_function()
2466 p_mem_node->length, p_mem_node->next); configure_new_function()
2514 bus_node->next = NULL; configure_new_function()
2519 io_node->next = NULL; configure_new_function()
2532 mem_node->next = NULL; configure_new_function()
2542 p_mem_node->next = NULL; configure_new_function()
2612 hold_bus_node->next = func->bus_head; configure_new_function()
2654 hold_IO_node->next = func->io_head; configure_new_function()
2671 hold_IO_node->next = func->io_head; configure_new_function()
2676 hold_IO_node->next = func->io_head; configure_new_function()
2704 hold_mem_node->next = func->mem_head; configure_new_function()
2723 hold_mem_node->next = func->mem_head; configure_new_function()
2728 hold_mem_node->next = func->mem_head; configure_new_function()
2757 hold_p_mem_node->next = func->p_mem_head; configure_new_function()
2774 hold_p_mem_node->next = func->p_mem_head; configure_new_function()
2779 hold_p_mem_node->next = func->p_mem_head; configure_new_function()
2828 dbg("Got io_node start = %8.8x, length = %8.8x next (%p)\n", configure_new_function()
2829 io_node->base, io_node->length, io_node->next); configure_new_function()
2836 io_node->next = func->io_head; configure_new_function()
2852 p_mem_node->next = func->p_mem_head; configure_new_function()
2868 mem_node->next = func->mem_head; configure_new_function()
H A Dcpqphp_pci.c433 /* this loop skips to the next present function cpqhp_save_config()
523 /* this loop skips to the next present function cpqhp_save_slot_config()
562 struct pci_func *next; cpqhp_save_base_addr_length() local
581 next = cpqhp_slot_list[sub_bus]; cpqhp_save_base_addr_length()
583 while (next != NULL) { cpqhp_save_base_addr_length()
584 rc = cpqhp_save_base_addr_length(ctrl, next); cpqhp_save_base_addr_length()
588 next = next->next; cpqhp_save_base_addr_length()
672 /* find the next device in this slot */ cpqhp_save_base_addr_length()
742 bus_node->next = func->bus_head; cpqhp_save_used_resources()
757 io_node->next = func->io_head; cpqhp_save_used_resources()
773 mem_node->next = func->mem_head; cpqhp_save_used_resources()
789 p_mem_node->next = func->p_mem_head; cpqhp_save_used_resources()
822 io_node->next = func->io_head; cpqhp_save_used_resources()
839 p_mem_node->next = func->p_mem_head; cpqhp_save_used_resources()
856 mem_node->next = func->mem_head; cpqhp_save_used_resources()
893 io_node->next = func->io_head; cpqhp_save_used_resources()
910 p_mem_node->next = func->p_mem_head; cpqhp_save_used_resources()
927 mem_node->next = func->mem_head; cpqhp_save_used_resources()
935 /* find the next device in this slot */ cpqhp_save_used_resources()
958 struct pci_func *next; cpqhp_configure_board() local
985 next = cpqhp_slot_list[sub_bus]; cpqhp_configure_board()
987 while (next != NULL) { cpqhp_configure_board()
988 rc = cpqhp_configure_board(ctrl, next); cpqhp_configure_board()
992 next = next->next; cpqhp_configure_board()
1039 struct pci_func *next; cpqhp_valid_replace() local
1083 next = cpqhp_slot_list[secondary_bus]; cpqhp_valid_replace()
1085 while (next != NULL) { cpqhp_valid_replace()
1086 rc = cpqhp_valid_replace(ctrl, next); cpqhp_valid_replace()
1090 next = next->next; cpqhp_valid_replace()
1154 /* Get the next function */ cpqhp_valid_replace()
1318 io_node->next = ctrl->io_head; cpqhp_find_available_resources()
1321 io_node->next = func->io_head; cpqhp_find_available_resources()
1341 mem_node->next = ctrl->mem_head; cpqhp_find_available_resources()
1344 mem_node->next = func->mem_head; cpqhp_find_available_resources()
1366 p_mem_node->next = ctrl->p_mem_head; cpqhp_find_available_resources()
1369 p_mem_node->next = func->p_mem_head; cpqhp_find_available_resources()
1389 bus_node->next = ctrl->bus_head; cpqhp_find_available_resources()
1392 bus_node->next = func->bus_head; cpqhp_find_available_resources()
1435 t_node = node->next; cpqhp_return_board_resources()
1443 t_node = node->next; cpqhp_return_board_resources()
1451 t_node = node->next; cpqhp_return_board_resources()
1459 t_node = node->next; cpqhp_return_board_resources()
1487 res = res->next; cpqhp_destroy_resource_list()
1496 res = res->next; cpqhp_destroy_resource_list()
1505 res = res->next; cpqhp_destroy_resource_list()
1514 res = res->next; cpqhp_destroy_resource_list()
1534 res = res->next; cpqhp_destroy_board_resources()
1543 res = res->next; cpqhp_destroy_board_resources()
1552 res = res->next; cpqhp_destroy_board_resources()
1561 res = res->next; cpqhp_destroy_board_resources()
H A Dibmphp_res.c417 range_cur = range_cur->next; add_bus_range()
433 range->next = range_cur; add_bus_range()
438 range->next = NULL; add_bus_range()
439 range_prev->next = range; add_bus_range()
444 range_prev->next = range; add_bus_range()
445 range->next = range_cur; add_bus_range()
452 range_cur = range_cur->next; add_bus_range()
489 if (res->next) update_resources()
490 res = res->next; update_resources()
503 res = res->next; update_resources()
543 range = range->next; fix_me()
546 if (res->next) fix_me()
547 res = res->next; fix_me()
591 * the ptrs to next and nextRange if needed.
643 range_cur = range_cur->next; ibmphp_add_resource()
680 res->next = NULL; ibmphp_add_resource()
692 if (res_cur->next) ibmphp_add_resource()
693 res_cur = res_cur->next; ibmphp_add_resource()
702 res->next = NULL; ibmphp_add_resource()
710 res_cur = res_cur->next; ibmphp_add_resource()
714 res_prev->next = res; ibmphp_add_resource()
715 res->next = NULL; ibmphp_add_resource()
733 res_prev->next = res; ibmphp_add_resource()
737 res->next = res_cur; ibmphp_add_resource()
744 res->next = NULL; ibmphp_add_resource()
762 res->next = NULL; ibmphp_add_resource()
822 if (res_cur->next) ibmphp_remove_resource()
823 res_cur = res_cur->next; ibmphp_remove_resource()
845 if (mem_cur->next) ibmphp_remove_resource()
846 mem_cur = mem_cur->next; ibmphp_remove_resource()
857 bus_cur->firstPFMemFromMem = res_cur->next; ibmphp_remove_resource()
859 res_prev->next = res_cur->next; ibmphp_remove_resource()
864 if (res_cur->next) ibmphp_remove_resource()
865 res_cur = res_cur->next; ibmphp_remove_resource()
880 if (res_cur->next) { ibmphp_remove_resource()
883 bus_cur->firstIO = res_cur->next; ibmphp_remove_resource()
886 bus_cur->firstMem = res_cur->next; ibmphp_remove_resource()
889 bus_cur->firstPFMem = res_cur->next; ibmphp_remove_resource()
920 if (res_cur->next) { ibmphp_remove_resource()
922 res_prev->next = res_cur->next; ibmphp_remove_resource()
924 res_prev->nextRange = res_cur->next; ibmphp_remove_resource()
926 res_prev->next = NULL; ibmphp_remove_resource()
929 res_prev->next = NULL; ibmphp_remove_resource()
960 range = range->next; find_range()
1084 if (!res_cur->next) { ibmphp_check_resource()
1203 if (res_cur->next) ibmphp_check_resource()
1204 res_cur = res_cur->next; ibmphp_check_resource()
1260 range = range->next; ibmphp_check_resource()
1326 range = range->next; ibmphp_check_resource()
1386 if (res_cur->next) ibmphp_remove_bus()
1387 res_cur = res_cur->next; ibmphp_remove_bus()
1399 if (res_cur->next) ibmphp_remove_bus()
1400 res_cur = res_cur->next; ibmphp_remove_bus()
1412 if (res_cur->next) ibmphp_remove_bus()
1413 res_cur = res_cur->next; ibmphp_remove_bus()
1426 res_cur = res_cur->next; ibmphp_remove_bus()
1460 range_cur = range_cur->next; remove_ranges()
1474 range_cur = range_cur->next; remove_ranges()
1488 range_cur = range_cur->next; remove_ranges()
1534 if (res_cur->next) ibmphp_find_resource()
1535 res_cur = res_cur->next; ibmphp_find_resource()
1548 res_cur = res_cur->next; ibmphp_find_resource()
1581 struct list_head *next; ibmphp_free_resources() local
1585 list_for_each_safe (tmp, next, &gbuses) { ibmphp_free_resources()
1593 range_cur = range_cur->next; ibmphp_free_resources()
1604 range_cur = range_cur->next; ibmphp_free_resources()
1615 range_cur = range_cur->next; ibmphp_free_resources()
1625 if (res_cur->next) ibmphp_free_resources()
1626 res_cur = res_cur->next; ibmphp_free_resources()
1638 if (res_cur->next) ibmphp_free_resources()
1639 res_cur = res_cur->next; ibmphp_free_resources()
1651 if (res_cur->next) ibmphp_free_resources()
1652 res_cur = res_cur->next; ibmphp_free_resources()
1665 res_cur = res_cur->next; ibmphp_free_resources()
1699 for (pfmem_cur = bus_cur->firstPFMem, pfmem_prev = NULL; pfmem_cur; pfmem_prev = pfmem_cur, pfmem_cur = pfmem_cur->next) { once_over()
1702 pfmem_prev->next = pfmem_cur->next; once_over()
1704 bus_cur->firstPFMem = pfmem_cur->next; once_over()
1707 pfmem_cur->next = NULL; once_over()
1713 pfmem_cur->next = bus_cur->firstPFMemFromMem; once_over()
1747 pfmem->next = bus_cur->firstPFMemFromMem; ibmphp_add_pfmem_from_mem()
1749 pfmem->next = NULL; ibmphp_add_pfmem_from_mem()
1812 range = range->next; ibmphp_print_test()
1822 range = range->next; ibmphp_print_test()
1833 range = range->next; ibmphp_print_test()
1846 if (res->next) ibmphp_print_test()
1847 res = res->next; ibmphp_print_test()
1861 if (res->next) ibmphp_print_test()
1862 res = res->next; ibmphp_print_test()
1876 if (res->next) ibmphp_print_test()
1877 res = res->next; ibmphp_print_test()
1892 res = res->next; ibmphp_print_test()
1920 range_cur = range_cur->next; range_exists_already()
/linux-4.4.14/drivers/staging/lustre/lustre/lov/
H A Dlovsub_dev.c94 const char *name, struct lu_device *next) lovsub_device_init()
100 next->ld_site = d->ld_site; lovsub_device_init()
101 ldt = next->ld_type; lovsub_device_init()
103 rc = ldt->ldt_ops->ldto_device_init(env, next, ldt->ldt_name, NULL); lovsub_device_init()
105 next->ld_site = NULL; lovsub_device_init()
109 lu_device_get(next); lovsub_device_init()
110 lu_ref_add(&next->ld_reference, "lu-stack", &lu_site_init); lovsub_device_init()
111 lsd->acid_next = lu2cl_dev(next); lovsub_device_init()
118 struct lu_device *next; lovsub_device_fini() local
122 next = cl2lu_dev(lsd->acid_next); lovsub_device_fini()
125 return next; lovsub_device_fini()
132 struct lu_device *next = cl2lu_dev(lsd->acid_next); lovsub_device_free() local
140 return next; lovsub_device_free()
93 lovsub_device_init(const struct lu_env *env, struct lu_device *d, const char *name, struct lu_device *next) lovsub_device_init() argument
/linux-4.4.14/drivers/scsi/arm/
H A Dmsgqueue.c30 msgq->free = mq->next; mqe_alloc()
44 mq->next = msgq->free; mqe_free()
62 msgq->entries[i].next = &msgq->entries[i + 1]; msgqueue_initialise()
64 msgq->entries[NR_MESSAGES - 1].next = NULL; msgqueue_initialise()
88 for (mq = msgq->qe; mq; mq = mq->next) msgqueue_msglength()
105 for (mq = msgq->qe; mq && msgno; mq = mq->next, msgno--); msgqueue_getmsg()
134 mq->next = NULL; msgqueue_addmsg()
138 mqp = &(*mqp)->next; msgqueue_addmsg()
156 mqnext = mq->next; msgqueue_flush()
/linux-4.4.14/drivers/media/pci/zoran/
H A Dvideocodec.c60 struct attached_list *next; member in struct:attached_list
67 struct codec_list *next; member in struct:codec_list
141 while (a->next) videocodec_attach()
142 a = a->next; // find end videocodec_attach()
143 a->next = ptr; videocodec_attach()
155 h = h->next; videocodec_attach()
209 h->list = a->next; videocodec_detach()
213 prev->next = a->next; videocodec_detach()
224 a = a->next; videocodec_detach()
226 h = h->next; videocodec_detach()
258 while (h->next) videocodec_register()
259 h = h->next; // find the end videocodec_register()
260 h->next = ptr; videocodec_register()
301 codeclist_top = h->next; videocodec_unregister()
305 prev->next = h->next; videocodec_unregister()
313 h = h->next; videocodec_unregister()
344 a = a->next; proc_videocodecs_show()
346 h = h->next; proc_videocodecs_show()
/linux-4.4.14/arch/um/include/asm/
H A Dmmu_context.h50 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
55 if(prev != next){ switch_mm()
57 cpumask_set_cpu(cpu, mm_cpumask(next)); switch_mm()
58 if(next != &init_mm) switch_mm()
59 __switch_mm(&next->context.id); switch_mm()
/linux-4.4.14/sound/isa/gus/
H A Dgus_mem.c55 nblock->next = pblock; snd_gf1_mem_xalloc()
60 nblock->prev->next = nblock; snd_gf1_mem_xalloc()
64 pblock = pblock->next; snd_gf1_mem_xalloc()
66 nblock->next = NULL; snd_gf1_mem_xalloc()
72 alloc->last->next = nblock; snd_gf1_mem_xalloc()
86 alloc->first = block->next; snd_gf1_mem_xfree()
87 if (block->next) snd_gf1_mem_xfree()
88 block->next->prev = NULL; snd_gf1_mem_xfree()
90 block->prev->next = block->next; snd_gf1_mem_xfree()
91 if (block->next) snd_gf1_mem_xfree()
92 block->next->prev = block->prev; snd_gf1_mem_xfree()
97 block->prev->next = NULL; snd_gf1_mem_xfree()
99 block->next->prev = block->prev; snd_gf1_mem_xfree()
101 block->prev->next = block->next; snd_gf1_mem_xfree()
113 for (block = alloc->first; block; block = block->next) { snd_gf1_mem_look()
129 for (block = alloc->first; block; block = block->next) snd_gf1_mem_share()
154 block->prev = block->next = NULL; snd_gf1_mem_find()
155 for (pblock = alloc->first, idx = 0; pblock; pblock = pblock->next) { snd_gf1_mem_find()
161 if (pblock->next) { snd_gf1_mem_find()
162 if (pblock->ptr + pblock->size == pblock->next->ptr) snd_gf1_mem_find()
164 if (pblock->next->ptr < boundary) snd_gf1_mem_find()
165 ptr2 = pblock->next->ptr; snd_gf1_mem_find()
280 nblock = block->next; snd_gf1_mem_done()
311 for (block = alloc->first, i = 0; block; block = block->next, i++) { snd_gf1_mem_info_read()
/linux-4.4.14/drivers/isdn/i4l/
H A Disdn_net.h87 nd->queue = nd->queue->next; isdn_net_get_locked_lp()
94 nd->queue = nd->queue->next; isdn_net_get_locked_lp()
118 lp->last->next = nlp; isdn_net_add_to_bundle()
120 nlp->next = lp; isdn_net_add_to_bundle()
136 // printk(KERN_DEBUG "%s: lp:%s(%p) mlp:%s(%p) last(%p) next(%p) mndq(%p)\n", isdn_net_rm_from_bundle()
137 // __func__, lp->name, lp, master_lp->name, master_lp, lp->last, lp->next, master_lp->netdev->queue); isdn_net_rm_from_bundle()
139 lp->last->next = lp->next; isdn_net_rm_from_bundle()
140 lp->next->last = lp->last; isdn_net_rm_from_bundle()
142 master_lp->netdev->queue = lp->next; isdn_net_rm_from_bundle()
143 if (lp->next == lp) { /* last in queue */ isdn_net_rm_from_bundle()
147 lp->next = lp->last = lp; /* (re)set own pointers */ isdn_net_rm_from_bundle()
/linux-4.4.14/mm/kasan/
H A Dkasan_init.c66 unsigned long next; zero_pmd_populate() local
69 next = pmd_addr_end(addr, end); zero_pmd_populate()
80 zero_pte_populate(pmd, addr, next); zero_pmd_populate()
81 } while (pmd++, addr = next, addr != end); zero_pmd_populate()
88 unsigned long next; zero_pud_populate() local
91 next = pud_addr_end(addr, end); zero_pud_populate()
105 zero_pmd_populate(pud, addr, next); zero_pud_populate()
106 } while (pud++, addr = next, addr != end); zero_pud_populate()
121 unsigned long next; kasan_populate_zero_shadow() local
124 next = pgd_addr_end(addr, end); kasan_populate_zero_shadow()
150 zero_pud_populate(pgd, addr, next); kasan_populate_zero_shadow()
151 } while (pgd++, addr = next, addr != end); kasan_populate_zero_shadow()
/linux-4.4.14/drivers/isdn/hardware/eicon/
H A Dpr_pc.h27 word NextReq; /* pointer to next Req Buffer */
28 word NextRc; /* pointer to next Rc Buffer */
29 word NextInd; /* pointer to next Ind Buffer */
46 word next; member in struct:__anon5602
56 word next; member in struct:__anon5603
65 word next; member in struct:__anon5604
/linux-4.4.14/drivers/acpi/
H A Dwakeup.c32 struct list_head *node, *next; acpi_enable_wakeup_devices() local
34 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_enable_wakeup_devices()
59 struct list_head *node, *next; acpi_disable_wakeup_devices() local
61 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_disable_wakeup_devices()
81 struct list_head *node, *next; acpi_wakeup_device_init() local
84 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_wakeup_device_init()
H A Dproc.c24 struct list_head *node, *next; acpi_system_wakeup_device_seq_show() local
29 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_system_wakeup_device_seq_show()
57 dev->physical_node_list.next) acpi_system_wakeup_device_seq_show()
98 struct list_head *node, *next; acpi_system_write_wakeup_device() local
111 list_for_each_safe(node, next, &acpi_wakeup_device_list) { acpi_system_write_wakeup_device()
/linux-4.4.14/fs/f2fs/
H A Dshrinker.c46 p = f2fs_list.next; f2fs_shrink_count()
52 p = p->next; f2fs_shrink_count()
67 p = p->next; f2fs_shrink_count()
87 p = f2fs_list.next; f2fs_shrink_scan()
96 p = p->next; f2fs_shrink_scan()
115 p = p->next; f2fs_shrink_scan()
/linux-4.4.14/drivers/acpi/acpica/
H A Dpstree.c107 arg = arg->common.next; acpi_ps_get_arg()
166 while (prev_arg->common.next) { acpi_ps_append_arg()
167 prev_arg = prev_arg->common.next; acpi_ps_append_arg()
169 prev_arg->common.next = arg; acpi_ps_append_arg()
180 arg = arg->common.next; acpi_ps_append_arg()
195 * DESCRIPTION: Get next op in tree (walking the tree in depth-first order)
203 union acpi_parse_object *next = NULL; acpi_ps_get_depth_next() local
215 next = acpi_ps_get_arg(op, 0); acpi_ps_get_depth_next()
216 if (next) { acpi_ps_get_depth_next()
217 return (next); acpi_ps_get_depth_next()
222 next = op->common.next; acpi_ps_get_depth_next()
223 if (next) { acpi_ps_get_depth_next()
224 return (next); acpi_ps_get_depth_next()
234 arg = arg->common.next; acpi_ps_get_depth_next()
244 if (parent->common.next) { acpi_ps_get_depth_next()
248 return (parent->common.next); acpi_ps_get_depth_next()
255 return (next); acpi_ps_get_depth_next()
H A Dpswalk.c65 union acpi_parse_object *next = NULL; acpi_ps_delete_parse_tree() local
80 next = acpi_ps_get_arg(op, 0); acpi_ps_delete_parse_tree()
81 if (next) { acpi_ps_delete_parse_tree()
85 op = next; acpi_ps_delete_parse_tree()
92 next = op->common.next; acpi_ps_delete_parse_tree()
102 if (next) { acpi_ps_delete_parse_tree()
103 op = next; acpi_ps_delete_parse_tree()
H A Devgpeutil.c98 gpe_block = gpe_block->next; acpi_ev_walk_gpe_list()
101 gpe_xrupt_info = gpe_xrupt_info->next; acpi_ev_walk_gpe_list()
184 next_gpe_xrupt = next_gpe_xrupt->next; acpi_ev_get_gpe_xrupt_block()
201 while (next_gpe_xrupt->next) { acpi_ev_get_gpe_xrupt_block()
202 next_gpe_xrupt = next_gpe_xrupt->next; acpi_ev_get_gpe_xrupt_block()
205 next_gpe_xrupt->next = gpe_xrupt; acpi_ev_get_gpe_xrupt_block()
271 gpe_xrupt->previous->next = gpe_xrupt->next; acpi_ev_delete_gpe_xrupt()
275 acpi_gbl_gpe_xrupt_list_head = gpe_xrupt->next; acpi_ev_delete_gpe_xrupt()
278 if (gpe_xrupt->next) { acpi_ev_delete_gpe_xrupt()
279 gpe_xrupt->next->previous = gpe_xrupt->previous; acpi_ev_delete_gpe_xrupt()
310 struct acpi_gpe_notify_info *next; acpi_ev_delete_gpe_handlers() local
345 next = notify->next; acpi_ev_delete_gpe_handlers()
347 notify = next; acpi_ev_delete_gpe_handlers()
H A Dexdump.c156 {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(region.next), "Next"}
244 {ACPI_EXD_HDLR_LIST, ACPI_EXD_OFFSET(address_space.next), "Next"},
257 {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[0]),
259 {ACPI_EXD_POINTER, ACPI_EXD_OFFSET(notify.next[1]), "Next Device Notify"}
373 union acpi_operand_object *next; acpi_ex_dump_object() local
471 next = start; acpi_ex_dump_object()
473 acpi_os_printf("%20s : %p", name, next); acpi_ex_dump_object()
474 if (next) { acpi_ex_dump_object()
477 (next), next->common.type); acpi_ex_dump_object()
479 while (next->common.next_object) { acpi_ex_dump_object()
480 if ((next->common.type == acpi_ex_dump_object()
482 data = next; acpi_ex_dump_object()
485 next = next->common.next_object; acpi_ex_dump_object()
486 acpi_os_printf("->%p(%s %2.2X)", next, acpi_ex_dump_object()
488 (next), acpi_ex_dump_object()
489 next->common.type); acpi_ex_dump_object()
491 if ((next == start) || (next == data)) { acpi_ex_dump_object()
505 next = start; acpi_ex_dump_object()
507 acpi_os_printf("%20s : %p", name, next); acpi_ex_dump_object()
508 if (next) { acpi_ex_dump_object()
511 (next), next->common.type); acpi_ex_dump_object()
513 while (next->address_space.next) { acpi_ex_dump_object()
514 if ((next->common.type == acpi_ex_dump_object()
516 data = next; acpi_ex_dump_object()
519 next = next->address_space.next; acpi_ex_dump_object()
520 acpi_os_printf("->%p(%s %2.2X)", next, acpi_ex_dump_object()
522 (next), acpi_ex_dump_object()
523 next->common.type); acpi_ex_dump_object()
525 if ((next == start) || (next == data)) { acpi_ex_dump_object()
539 next = start; acpi_ex_dump_object()
541 acpi_os_printf("%20s : %p", name, next); acpi_ex_dump_object()
542 if (next) { acpi_ex_dump_object()
545 (next), next->common.type); acpi_ex_dump_object()
547 while (next->region.next) { acpi_ex_dump_object()
548 if ((next->common.type == acpi_ex_dump_object()
550 data = next; acpi_ex_dump_object()
553 next = next->region.next; acpi_ex_dump_object()
554 acpi_os_printf("->%p(%s %2.2X)", next, acpi_ex_dump_object()
556 (next), acpi_ex_dump_object()
557 next->common.type); acpi_ex_dump_object()
559 if ((next == start) || (next == data)) { acpi_ex_dump_object()
H A Dutaddress.c106 range_info->next = acpi_gbl_address_range_list[space_id]; acpi_ut_add_address_range()
157 range_info->next; acpi_ut_remove_address_range()
159 prev->next = range_info->next; acpi_ut_remove_address_range()
176 range_info = range_info->next; acpi_ut_remove_address_range()
259 range_info = range_info->next; acpi_ut_check_address_range()
280 struct acpi_address_range *next; acpi_ut_delete_address_lists() local
287 next = acpi_gbl_address_range_list[i]; acpi_ut_delete_address_lists()
289 while (next) { acpi_ut_delete_address_lists()
290 range_info = next; acpi_ut_delete_address_lists()
291 next = range_info->next; acpi_ut_delete_address_lists()
/linux-4.4.14/arch/alpha/kernel/
H A Derr_common.c151 /* return the next header */ el_process_header_subpacket()
160 struct el_subpacket *next = NULL; el_process_subpacket_reg()
163 for (; h && h->class != header->class; h = h->next); el_process_subpacket_reg()
164 if (h) next = h->handler(header); el_process_subpacket_reg()
166 return next;
198 struct el_subpacket *next = NULL; el_process_subpacket()
205 next = el_process_header_subpacket(header); el_process_subpacket()
208 if (NULL == (next = el_process_subpacket_reg(header))) { el_process_subpacket()
217 return next;
226 for (a = subpacket_annotation_list; a; a = a->next) { el_annotate_subpacket()
287 for (; a->next != NULL; a = a->next) { cdl_register_subpacket_annotation()
295 a->next = new; cdl_register_subpacket_annotation()
297 new->next = NULL; cdl_register_subpacket_annotation()
309 for (; h->next != NULL; h = h->next) { cdl_register_subpacket_handler()
316 h->next = new; cdl_register_subpacket_handler()
318 new->next = NULL; cdl_register_subpacket_handler()
158 struct el_subpacket *next = NULL; el_process_subpacket_reg() local
196 struct el_subpacket *next = NULL; el_process_subpacket() local
H A Dgct.c39 if (node->next) gct6_find_nodes()
40 status |= gct6_find_nodes(GCT_NODE_PTR(node->next), search); gct6_find_nodes()
/linux-4.4.14/kernel/
H A Dtask_work.c5 static struct callback_head work_exited; /* all we need is ->next == NULL */
35 work->next = head; task_work_add()
70 pprev = &work->next; task_work_cancel()
71 else if (cmpxchg(pprev, work, work->next) == work) task_work_cancel()
90 struct callback_head *work, *head, *next; task_work_run() local
114 next = work->next; task_work_run()
116 work = next; task_work_run()
/linux-4.4.14/arch/mips/mm/
H A Dgup.c104 unsigned long next; gup_pmd_range() local
111 next = pmd_addr_end(addr, end); gup_pmd_range()
126 if (!gup_huge_pmd(pmd, addr, next, write, pages,nr)) gup_pmd_range()
129 if (!gup_pte_range(pmd, addr, next, write, pages,nr)) gup_pmd_range()
132 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
170 unsigned long next; gup_pud_range() local
177 next = pud_addr_end(addr, end); gup_pud_range()
181 if (!gup_huge_pud(pud, addr, next, write, pages,nr)) gup_pud_range()
184 if (!gup_pmd_range(pud, addr, next, write, pages,nr)) gup_pud_range()
187 } while (pudp++, addr = next, addr != end); gup_pud_range()
201 unsigned long next; __get_user_pages_fast() local
236 next = pgd_addr_end(addr, end); __get_user_pages_fast()
239 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) __get_user_pages_fast()
241 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
268 unsigned long next; get_user_pages_fast() local
286 next = pgd_addr_end(addr, end); get_user_pages_fast()
289 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) get_user_pages_fast()
291 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
/linux-4.4.14/arch/nios2/include/asm/
H A Dswitch_to.h17 #define switch_to(prev, next, last) \
26 : "r" (prev), "r" (next) \
/linux-4.4.14/arch/x86/mm/
H A Dinit_64.c73 unsigned long next; ident_pud_init() local
75 for (; addr < end; addr = next) { ident_pud_init()
79 next = (addr & PUD_MASK) + PUD_SIZE; ident_pud_init()
80 if (next > end) ident_pud_init()
81 next = end; ident_pud_init()
85 ident_pmd_init(info->pmd_flag, pmd, addr, next); ident_pud_init()
91 ident_pmd_init(info->pmd_flag, pmd, addr, next); ident_pud_init()
101 unsigned long next; kernel_ident_mapping_init() local
105 for (; addr < end; addr = next) { kernel_ident_mapping_init()
109 next = (addr & PGDIR_MASK) + PGDIR_SIZE; kernel_ident_mapping_init()
110 if (next > end) kernel_ident_mapping_init()
111 next = end; kernel_ident_mapping_init()
115 result = ident_pud_init(info, pud, addr, next); kernel_ident_mapping_init()
124 result = ident_pud_init(info, pud, addr, next); kernel_ident_mapping_init()
405 unsigned long pages = 0, next; phys_pte_init() local
411 for (i = pte_index(addr); i < PTRS_PER_PTE; i++, addr = next, pte++) { phys_pte_init()
412 next = (addr & PAGE_MASK) + PAGE_SIZE; phys_pte_init()
415 !e820_any_mapped(addr & PAGE_MASK, next, E820_RAM) && phys_pte_init()
416 !e820_any_mapped(addr & PAGE_MASK, next, E820_RESERVED_KERN)) phys_pte_init()
450 unsigned long pages = 0, next; phys_pmd_init() local
455 for (; i < PTRS_PER_PMD; i++, address = next) { phys_pmd_init()
460 next = (address & PMD_MASK) + PMD_SIZE; phys_pmd_init()
463 !e820_any_mapped(address & PMD_MASK, next, E820_RAM) && phys_pmd_init()
464 !e820_any_mapped(address & PMD_MASK, next, E820_RESERVED_KERN)) phys_pmd_init()
493 last_map_addr = next; phys_pmd_init()
506 last_map_addr = next; phys_pmd_init()
525 unsigned long pages = 0, next; phys_pud_init() local
529 for (; i < PTRS_PER_PUD; i++, addr = next) { phys_pud_init()
534 next = (addr & PUD_MASK) + PUD_SIZE; phys_pud_init()
537 !e820_any_mapped(addr & PUD_MASK, next, E820_RAM) && phys_pud_init()
538 !e820_any_mapped(addr & PUD_MASK, next, E820_RESERVED_KERN)) phys_pud_init()
566 last_map_addr = next; phys_pud_init()
579 last_map_addr = next; phys_pud_init()
604 unsigned long next, last_map_addr = end; kernel_physical_mapping_init() local
611 for (; start < end; start = next) { kernel_physical_mapping_init()
615 next = (start & PGDIR_MASK) + PGDIR_SIZE; kernel_physical_mapping_init()
722 magic = (unsigned long)page->lru.next; free_pagetable()
794 unsigned long next, pages = 0; remove_pte_table() local
800 for (; addr < end; addr = next, pte++) { remove_pte_table()
801 next = (addr + PAGE_SIZE) & PAGE_MASK; remove_pte_table()
802 if (next > end) remove_pte_table()
803 next = end; remove_pte_table()
818 IS_ALIGNED(next, PAGE_SIZE)) { remove_pte_table()
843 memset((void *)addr, PAGE_INUSE, next - addr); remove_pte_table()
866 unsigned long next, pages = 0; remove_pmd_table() local
872 for (; addr < end; addr = next, pmd++) { remove_pmd_table()
873 next = pmd_addr_end(addr, end); remove_pmd_table()
880 IS_ALIGNED(next, PMD_SIZE)) { remove_pmd_table()
891 memset((void *)addr, PAGE_INUSE, next - addr); remove_pmd_table()
909 remove_pte_table(pte_base, addr, next, direct); remove_pmd_table()
922 unsigned long next, pages = 0; remove_pud_table() local
928 for (; addr < end; addr = next, pud++) { remove_pud_table()
929 next = pud_addr_end(addr, end); remove_pud_table()
936 IS_ALIGNED(next, PUD_SIZE)) { remove_pud_table()
947 memset((void *)addr, PAGE_INUSE, next - addr); remove_pud_table()
965 remove_pmd_table(pmd_base, addr, next, direct); remove_pud_table()
977 unsigned long next; remove_pagetable() local
983 for (addr = start; addr < end; addr = next) { remove_pagetable()
984 next = pgd_addr_end(addr, end); remove_pagetable()
991 remove_pud_table(pud, addr, next, direct); remove_pagetable()
1242 unsigned long next; vmemmap_populate_hugepages() local
1247 for (addr = start; addr < end; addr = next) { vmemmap_populate_hugepages()
1248 next = pmd_addr_end(addr, end); vmemmap_populate_hugepages()
1285 vmemmap_verify((pte_t *)pmd, node, addr, next); vmemmap_populate_hugepages()
1289 if (vmemmap_populate_basepages(addr, next, node)) vmemmap_populate_hugepages()
1314 unsigned long next; register_page_bootmem_memmap() local
1321 for (; addr < end; addr = next) { register_page_bootmem_memmap()
1326 next = (addr + PAGE_SIZE) & PAGE_MASK; register_page_bootmem_memmap()
1333 next = (addr + PAGE_SIZE) & PAGE_MASK; register_page_bootmem_memmap()
1339 next = (addr + PAGE_SIZE) & PAGE_MASK; register_page_bootmem_memmap()
1352 next = pmd_addr_end(addr, end); register_page_bootmem_memmap()
/linux-4.4.14/net/ipv4/
H A Dip_fragment.c314 struct sk_buff *xp = fp->next; ip_frag_reinit()
336 struct sk_buff *prev, *next; ip_frag_queue() local
405 next = NULL; ip_frag_queue()
409 for (next = qp->q.fragments; next != NULL; next = next->next) { ip_frag_queue()
410 if (FRAG_CB(next)->offset >= offset) ip_frag_queue()
412 prev = next; ip_frag_queue()
438 while (next && FRAG_CB(next)->offset < end) { ip_frag_queue()
439 int i = end - FRAG_CB(next)->offset; /* overlap is 'i' bytes */ ip_frag_queue()
441 if (i < next->len) { ip_frag_queue()
442 /* Eat head of the next overlapped fragment ip_frag_queue()
443 * and leave the loop. The next ones cannot overlap. ip_frag_queue()
445 if (!pskb_pull(next, i)) ip_frag_queue()
447 FRAG_CB(next)->offset += i; ip_frag_queue()
449 if (next->ip_summed != CHECKSUM_UNNECESSARY) ip_frag_queue()
450 next->ip_summed = CHECKSUM_NONE; ip_frag_queue()
453 struct sk_buff *free_it = next; ip_frag_queue()
458 next = next->next; ip_frag_queue()
461 prev->next = next; ip_frag_queue()
463 qp->q.fragments = next; ip_frag_queue()
474 skb->next = next; ip_frag_queue()
475 if (!next) ip_frag_queue()
478 prev->next = skb; ip_frag_queue()
544 head = prev->next; ip_frag_reasm()
549 fp->next = head->next; ip_frag_reasm()
550 if (!fp->next) ip_frag_reasm()
552 prev->next = fp; ip_frag_reasm()
555 head->next = qp->q.fragments->next; ip_frag_reasm()
586 clone->next = head->next; ip_frag_reasm()
587 head->next = clone; ip_frag_reasm()
600 skb_shinfo(head)->frag_list = head->next; ip_frag_reasm()
603 for (fp=head->next; fp; fp = fp->next) { ip_frag_reasm()
614 head->next = NULL; ip_frag_reasm()
/linux-4.4.14/arch/sparc/mm/
H A Dgup.c121 unsigned long next; gup_pmd_range() local
128 next = pmd_addr_end(addr, end); gup_pmd_range()
132 if (!gup_huge_pmd(pmdp, pmd, addr, next, gup_pmd_range()
135 } else if (!gup_pte_range(pmd, addr, next, write, gup_pmd_range()
138 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
146 unsigned long next; gup_pud_range() local
153 next = pud_addr_end(addr, end); gup_pud_range()
156 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) gup_pud_range()
158 } while (pudp++, addr = next, addr != end); gup_pud_range()
168 unsigned long next, flags; __get_user_pages_fast() local
182 next = pgd_addr_end(addr, end); __get_user_pages_fast()
185 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) __get_user_pages_fast()
187 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
198 unsigned long next; get_user_pages_fast() local
230 next = pgd_addr_end(addr, end); get_user_pages_fast()
233 if (!gup_pud_range(pgd, addr, next, write, pages, &nr)) get_user_pages_fast()
235 } while (pgdp++, addr = next, addr != end); get_user_pages_fast()
/linux-4.4.14/net/core/
H A Ddst.c70 struct dst_entry *dst, *next, head; dst_gc_task() local
74 next = dst_busy_list; dst_gc_task()
77 while ((dst = next) != NULL) { dst_gc_task()
78 next = dst->next; dst_gc_task()
79 prefetch(&next->next); dst_gc_task()
82 last->next = dst; dst_gc_task()
103 dst->next = next; dst_gc_task()
104 next = dst; dst_gc_task()
109 next = dst_garbage.list; dst_gc_task()
110 if (next) { dst_gc_task()
115 last->next = NULL; dst_gc_task()
116 dst_busy_list = head.next; dst_gc_task()
135 * if the next desired timer is more than 4 seconds in the dst_gc_task()
194 dst->next = NULL; dst_init()
236 dst->next = dst_garbage.list; __dst_free()
452 for (dst = dst_busy_list; dst; dst = dst->next) { dst_dev_event()
463 last->next = dst; dst_dev_event()
466 for (; dst; dst = dst->next) dst_dev_event()
/linux-4.4.14/drivers/scsi/pcmcia/
H A Dnsp_message.c22 * the next "msg in" if exists (no scsi phase changes). nsp_message_in()
45 /* catch a next signal */ nsp_message_in()
63 * the next "msg out" if exists (no scsi phase changes). nsp_message_out()
72 /* catch a next signal */ nsp_message_out()
/linux-4.4.14/drivers/md/
H A Draid0.h5 sector_t zone_end; /* Start of the next zone (in sectors) */
/linux-4.4.14/drivers/net/appletalk/
H A Dipddp.h20 struct ipddp_route *next; member in struct:ipddp_route
/linux-4.4.14/drivers/net/wireless/b43/
H A Ddebugfs.h58 /* The cached address for the next mmio16read file read */
60 /* The cached address for the next mmio32read file read */
63 /* The cached address for the next shm16read file read */
66 /* The cached address for the next shm32read file read */
/linux-4.4.14/arch/c6x/kernel/
H A Dswitch_to.S17 * struct thread_info *next,
23 || MV .L1X B4,A5 ; next
45 ;; Switch to next SP
61 B .S2 B3 ; return in next E1
/linux-4.4.14/include/net/
H A Datmclip.h33 struct clip_vcc *next; /* next VCC */ member in struct:clip_vcc
49 struct net_device *next; /* next CLIP interface */ member in struct:clip_priv
/linux-4.4.14/include/asm-generic/bitops/
H A Dfind.h6 * find_next_bit - find the next set bit in a memory region
11 * Returns the bit number for the next set bit
20 * find_next_zero_bit - find the next cleared bit in a memory region
25 * Returns the bit number of the next zero bit
/linux-4.4.14/arch/sh/include/cpu-sh5/cpu/
H A Dmmu_context.h16 /* This has to be a common function because the next location to fill
/linux-4.4.14/arch/m68k/include/asm/
H A Dswitch_to.h20 * Beware that resume now expects *next to be in d1 and the offset of
27 * pass prev in a0, next in a1
30 #define switch_to(prev,next,last) do { \
32 register void *_next __asm__ ("a1") = (next); \
/linux-4.4.14/arch/h8300/include/asm/
H A Dswitch_to.h20 * Beware that resume now expects *next to be in d1 and the offset of
27 * pass prev in a0, next in a1, offset of tss in d1, and whether
34 #define switch_to(prev, next, last) \
45 "r" (&(next->thread)), \
/linux-4.4.14/net/ieee802154/6lowpan/
H A Dreassembly.c132 struct sk_buff *prev, *next; lowpan_frag_queue() local
169 next = NULL; lowpan_frag_queue()
173 for (next = fq->q.fragments; next != NULL; next = next->next) { lowpan_frag_queue()
174 if (lowpan_802154_cb(next)->d_offset >= lowpan_frag_queue()
177 prev = next; lowpan_frag_queue()
182 skb->next = next; lowpan_frag_queue()
183 if (!next) lowpan_frag_queue()
186 prev->next = skb; lowpan_frag_queue()
236 head = prev->next; lowpan_frag_reasm()
242 fp->next = head->next; lowpan_frag_reasm()
243 if (!fp->next) lowpan_frag_reasm()
245 prev->next = fp; lowpan_frag_reasm()
248 head->next = fq->q.fragments->next; lowpan_frag_reasm()
269 clone->next = head->next; lowpan_frag_reasm()
270 head->next = clone; lowpan_frag_reasm()
285 for (fp = head->next; fp;) { lowpan_frag_reasm()
288 struct sk_buff *next = fp->next; lowpan_frag_reasm() local
300 fp = next; lowpan_frag_reasm()
304 head->next = NULL; lowpan_frag_reasm()
/linux-4.4.14/net/rose/
H A Drose_route.c70 rose_node = rose_node->next; rose_add_node()
84 rose_neigh = rose_neigh->next; rose_add_node()
129 rose_neigh->next = rose_neigh_list; rose_add_node()
147 rose_tmpn = rose_tmpn->next; rose_add_node()
169 rose_node->next = NULL; rose_add_node()
171 rose_tmpp->next = rose_node; rose_add_node()
172 rose_node->next = NULL; rose_add_node()
176 rose_node->next = rose_node_list; rose_add_node()
179 rose_tmpp->next = rose_node; rose_add_node()
180 rose_node->next = rose_tmpn; rose_add_node()
210 rose_node_list = rose_node->next; rose_remove_node()
215 while (s != NULL && s->next != NULL) { rose_remove_node()
216 if (s->next == rose_node) { rose_remove_node()
217 s->next = rose_node->next; rose_remove_node()
222 s = s->next; rose_remove_node()
239 rose_neigh_list = rose_neigh->next; rose_remove_neigh()
247 while (s != NULL && s->next != NULL) { rose_remove_neigh()
248 if (s->next == rose_neigh) { rose_remove_neigh()
249 s->next = rose_neigh->next; rose_remove_neigh()
257 s = s->next; rose_remove_neigh()
275 rose_route_list = rose_route->next; rose_remove_route()
280 while (s != NULL && s->next != NULL) { rose_remove_route()
281 if (s->next == rose_route) { rose_remove_route()
282 s->next = rose_route->next; rose_remove_route()
287 s = s->next; rose_remove_route()
311 rose_node = rose_node->next; rose_del_node()
325 rose_neigh = rose_neigh->next; rose_del_node()
397 sn->next = rose_neigh_list; rose_add_loopback_neigh()
418 rose_node = rose_node->next; rose_add_loopback_node()
436 rose_node->next = rose_node_list; rose_add_loopback_node()
462 rose_node = rose_node->next; rose_del_loopback_node()
490 rose_neigh = rose_neigh->next; rose_rt_device_down()
499 rose_node = rose_node->next; rose_rt_device_down()
539 rose_route = rose_route->next;
566 rose_node = rose_node->next; rose_clear_routes()
573 rose_neigh = rose_neigh->next; rose_clear_routes()
664 for (rose_route = rose_route_list; rose_route != NULL; rose_route = rose_route->next) rose_route_free_lci()
684 for (node = rose_node_list; node != NULL; node = node->next) { rose_get_neigh()
695 for (node = rose_node_list; node != NULL; node = node->next) { rose_get_neigh()
783 s = rose_route->next; rose_del_route_by_neigh()
801 rose_route = rose_route->next; rose_del_route_by_neigh()
820 rose_neigh = rose_neigh->next; rose_link_failed()
841 for (rose_neigh = rose_neigh_list; rose_neigh != NULL; rose_neigh = rose_neigh->next) { rose_link_device_down()
886 rose_neigh = rose_neigh->next; rose_route_frame()
955 * Route it to the next in line if we have an entry for it. rose_route_frame()
1001 rose_route = rose_route->next; rose_route_frame()
1035 rose_route = rose_route->next; rose_route_frame()
1066 rose_route->next = rose_route_list; rose_route_frame()
1096 rose_node = rose_node->next, ++i); __acquires()
1106 : ((struct rose_node *)v)->next; rose_node_next()
1146 .next = rose_node_next,
1175 rose_neigh = rose_neigh->next, ++i); __acquires()
1185 : ((struct rose_neigh *)v)->next; rose_neigh_next()
1230 .next = rose_neigh_next,
1260 rose_route = rose_route->next, ++i); __acquires()
1270 : ((struct rose_route *)v)->next; rose_route_next()
1316 .next = rose_route_next,
1347 rose_neigh = rose_neigh->next; rose_rt_free()
1354 rose_node = rose_node->next; rose_rt_free()
1361 rose_route = rose_route->next; rose_rt_free()
/linux-4.4.14/drivers/staging/rtl8192u/ieee80211/
H A Dieee80211_crypt.c45 for (ptr = ieee->crypt_deinit_list.next, n = ptr->next; ieee80211_crypt_deinit_entries()
46 ptr != &ieee->crypt_deinit_list; ptr = n, n = ptr->next) { ieee80211_crypt_deinit_entries()
136 for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) { ieee80211_unregister_crypto_ops()
167 for (ptr = hcrypt->algs.next; ptr != &hcrypt->algs; ptr = ptr->next) { ieee80211_get_crypto_ops()
229 for (ptr = hcrypt->algs.next, n = ptr->next; ptr != &hcrypt->algs; ieee80211_crypto_deinit()
230 ptr = n, n = ptr->next) { ieee80211_crypto_deinit()
/linux-4.4.14/drivers/staging/rdma/amso1100/
H A Dc2_alloc.c54 new_head->next = NULL; c2_alloc_mqsp_chunk()
57 /* build list where each index is the next free slot */ c2_alloc_mqsp_chunk()
79 struct sp_chunk *next; c2_free_mqsp_pool() local
82 next = root->next; c2_free_mqsp_pool()
85 root = next; c2_free_mqsp_pool()
99 } else if (head->next == NULL) { c2_alloc_mqsp()
100 if (c2_alloc_mqsp_chunk(c2dev, gfp_mask, &head->next) == c2_alloc_mqsp()
102 head = head->next; c2_alloc_mqsp()
109 head = head->next; c2_alloc_mqsp()
/linux-4.4.14/fs/nilfs2/
H A Dsegbuf.h42 * @next: Block number of the next full segment
54 sector_t next; member in struct:nilfs_segsum_info
63 * @sb_nextnum: Index number of the next full segment
100 #define NILFS_NEXT_SEGBUF(segbuf) NILFS_LIST_SEGBUF((segbuf)->sb_list.next)
103 #define NILFS_FIRST_SEGBUF(head) NILFS_LIST_SEGBUF((head)->next)
104 #define NILFS_SEGBUF_IS_LAST(segbuf, head) ((segbuf)->sb_list.next == (head))
111 (list_entry((head)->next, struct buffer_head, b_assoc_buffers))
113 (list_entry((bh)->b_assoc_buffers.next, struct buffer_head, \
115 #define NILFS_SEGBUF_BH_IS_LAST(bh, head) ((bh)->b_assoc_buffers.next == head)
/linux-4.4.14/arch/alpha/lib/
H A Dmemset.S34 bis $17,$1,$17 /* E0 (p-c latency, next cycle) */
35 sll $17,16,$1 /* E1 (p-c latency, next cycle) */
37 bis $17,$1,$17 /* E0 (p-c latency, next cycle) */
38 sll $17,32,$1 /* E1 (p-c latency, next cycle) */
39 bis $17,$1,$17 /* E0 (p-c latency, next cycle) */
/linux-4.4.14/drivers/isdn/divert/
H A Disdn_divert.c29 struct call_struc *next; /* pointer to next entry */ member in struct:call_struc
38 struct deflect_struc *next, *prev; member in struct:deflect_struc
48 static ulong next_id = 1; /* next info id */
94 cs->prev->next = cs->next; /* forward link */ deflect_timer_expire()
96 divert_head = cs->next; deflect_timer_expire()
97 if (cs->next) deflect_timer_expire()
98 cs->next->prev = cs->prev; /* back link */ deflect_timer_expire()
193 cs->next = divert_head; cf_command()
216 cs = cs->next; deflect_extern_action()
289 ds1 = ds1->next; insertrule()
296 ds->next = NULL; /* end of chain */ insertrule()
298 ds->prev->next = ds; /* last forward */ insertrule()
303 ds->next = ds1; /* next entry */ insertrule()
330 ds = ds->next; deleterule()
341 ds = ds->next; deleterule()
349 if (ds->next) deleterule()
350 ds->next->prev = ds->prev; /* backward chain */ deleterule()
355 ds->prev->next = ds->next; /* forward chain */ deleterule()
357 table_head = ds->next; /* start of chain */ deleterule()
377 ds = ds->next; getruleptr()
395 for (dv = table_head; dv; dv = dv->next) { isdn_divert_icall()
520 cs->next = divert_head; isdn_divert_icall()
543 cs = cs->next; deleteprocs()
744 cs = cs->next; prot_stat_callback()
757 cs1->prev->next = cs1->next; /* forward link */ prot_stat_callback()
759 divert_head = cs1->next; prot_stat_callback()
760 if (cs1->next) prot_stat_callback()
761 cs1->next->prev = cs1->prev; /* back link */ prot_stat_callback()
809 cs = cs->next; isdn_divert_stat_callback()
813 cs1->prev->next = cs1->next; /* forward link */ isdn_divert_stat_callback()
815 divert_head = cs1->next; isdn_divert_stat_callback()
816 if (cs1->next) isdn_divert_stat_callback()
817 cs1->next->prev = cs1->prev; /* back link */ isdn_divert_stat_callback()
/linux-4.4.14/fs/omfs/
H A Dfile.c37 u64 next, last; omfs_shrink_inode() local
45 next = inode->i_ino; omfs_shrink_inode()
52 bh = omfs_bread(inode->i_sb, next); omfs_shrink_inode()
61 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) omfs_shrink_inode()
69 last = next; omfs_shrink_inode()
70 next = be64_to_cpu(oe->e_next); omfs_shrink_inode()
89 if (next == ~0) omfs_shrink_inode()
92 bh = omfs_bread(inode->i_sb, next); omfs_shrink_inode()
141 /* trivially grow current extent, if next block is not taken */ omfs_grow_extent()
221 sector_t next, offset; omfs_get_block() local
239 next = inode->i_ino; omfs_get_block()
243 if (omfs_is_bad(sbi, (struct omfs_header *) bh->b_data, next)) omfs_get_block()
247 next = be64_to_cpu(oe->e_next); omfs_get_block()
262 if (next == ~0) omfs_get_block()
266 bh = omfs_bread(inode->i_sb, next); omfs_get_block()
/linux-4.4.14/drivers/char/agp/
H A Dfrontend.c59 curr = curr->next; agp_find_mem_by_key()
69 struct agp_memory *next; agp_remove_from_pool() local
75 next = temp->next; agp_remove_from_pool()
79 prev->next = next; agp_remove_from_pool()
80 if (next != NULL) agp_remove_from_pool()
81 next->prev = prev; agp_remove_from_pool()
85 if (next != NULL) agp_remove_from_pool()
86 next->prev = NULL; agp_remove_from_pool()
88 agp_fe.current_controller->pool = next; agp_remove_from_pool()
207 temp->next = prev; agp_insert_into_pool()
224 curr = curr->next; agp_find_private()
238 priv->next = prev; agp_insert_file_private()
244 struct agp_file_private *next; agp_remove_file_private() local
247 next = priv->next; agp_remove_file_private()
251 prev->next = next; agp_remove_file_private()
253 if (next != NULL) agp_remove_file_private()
254 next->prev = prev; agp_remove_file_private()
257 if (next != NULL) agp_remove_file_private()
258 next->prev = NULL; agp_remove_file_private()
260 agp_fe.file_priv_list = next; agp_remove_file_private()
302 controller = controller->next; agp_find_controller_by_pid()
325 controller->next = prev_controller; agp_insert_controller()
353 client = client->next; agp_remove_all_clients()
367 memory = memory->next; agp_remove_all_memory()
378 next_controller = controller->next; agp_remove_controller()
381 prev_controller->next = next_controller; agp_remove_controller()
419 clients = clients->next; agp_controller_make_current()
441 clients = clients->next; agp_controller_release_current()
467 client = client->next; agp_find_client_in_controller()
482 controller = controller->next; agp_find_controller_for_client()
504 client->next = prev_client; agp_insert_client()
542 next_client = client->next; agp_remove_client()
545 prev_client->next = next_client; agp_remove_client()
/linux-4.4.14/net/atm/
H A Dmpoa_caches.c47 entry = entry->next; in_cache_get()
68 entry = entry->next; in_cache_get_with_mask()
89 entry = entry->next; in_cache_get_by_vcc()
111 entry->next = client->in_cache; in_cache_add_entry()
201 entry->prev->next = entry->next; in_cache_remove_entry()
203 client->in_cache = entry->next; in_cache_remove_entry()
204 if (entry->next != NULL) in_cache_remove_entry()
205 entry->next->prev = entry->prev; in_cache_remove_entry()
237 next_entry = entry->next; clear_count_and_expired()
266 entry = entry->next; /* Entry in hold down */ check_resolving_entries()
279 entry = entry->next; check_resolving_entries()
294 entry = entry->next; check_resolving_entries()
320 entry = entry->next; refresh_entries()
346 entry = entry->next; eg_cache_get_by_cache_id()
367 entry = entry->next; eg_cache_get_by_tag()
389 entry = entry->next; eg_cache_get_by_vcc()
409 entry = entry->next; eg_cache_get_by_src_ip()
436 entry->prev->next = entry->next; eg_cache_remove_entry()
438 client->eg_cache = entry->next; eg_cache_remove_entry()
439 if (entry->next != NULL) eg_cache_remove_entry()
440 entry->next->prev = entry->prev; eg_cache_remove_entry()
474 entry->next = client->eg_cache; eg_cache_add_entry()
513 next_entry = entry->next; clear_expired()
/linux-4.4.14/scripts/dtc/
H A Dlivetree.c41 new->next = *labels; add_label()
79 assert(first->next == NULL); chain_property()
81 first->next = list; chain_property()
89 struct property *next; reverse_properties() local
92 next = p->next; reverse_properties()
93 p->next = head; reverse_properties()
95 p = next; reverse_properties()
154 new_node->proplist = new_prop->next; merge_nodes()
155 new_prop->next = NULL; merge_nodes()
231 prop->next = NULL; add_property()
235 p = &((*p)->next); add_property()
249 prop = prop->next; delete_property_by_name()
314 assert(first->next == NULL); chain_reserve_entry()
316 first->next = list; chain_reserve_entry()
325 new->next = NULL; add_reserve_entry()
330 for (last = list; last->next; last = last->next) add_reserve_entry()
333 last->next = new; add_reserve_entry()
602 ri = ri->next) sort_reserve_entries()
612 ri = ri->next) sort_reserve_entries()
619 tbl[i]->next = tbl[i+1]; sort_reserve_entries()
620 tbl[n-1]->next = NULL; sort_reserve_entries()
655 tbl[i]->next = tbl[i+1]; sort_properties()
656 tbl[n-1]->next = NULL; sort_properties()
/linux-4.4.14/arch/arm64/include/asm/
H A Dmmu_context.h32 static inline void contextidr_thread_switch(struct task_struct *next) contextidr_thread_switch() argument
38 : "r" (task_pid_nr(next))); contextidr_thread_switch()
41 static inline void contextidr_thread_switch(struct task_struct *next) contextidr_thread_switch() argument
127 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
132 if (prev == next) switch_mm()
139 if (next == &init_mm) { switch_mm()
144 check_and_switch_context(next, cpu); switch_mm()
148 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
H A Dspinlock_types.h29 u16 next; member in struct:__anon306
33 u16 next;
/linux-4.4.14/drivers/base/
H A Dmap.c21 struct probe *next; member in struct:kobj_map::probe
60 s = &(*s)->next; kobj_map()
61 p->next = *s; kobj_map()
81 for (s = &domain->probes[index % 255]; *s; s = &(*s)->next) { kobj_unmap()
84 *s = p->next; kobj_unmap()
103 for (p = domain->probes[MAJOR(dev) % 255]; p; p = p->next) { kobj_lookup()
/linux-4.4.14/fs/jffs2/
H A Dnodelist.c43 new->next = (*prev)->next; jffs2_add_fd_to_list()
53 prev = &((*prev)->next); jffs2_add_fd_to_list()
55 new->next = *prev; jffs2_add_fd_to_list()
73 struct jffs2_node_frag *next = frag_next(frag); jffs2_truncate_fragtree() local
77 frag = next; jffs2_truncate_fragtree()
384 or next node REF_NORMAL, as appropriate. */ jffs2_add_full_dnode_to_inode()
395 struct jffs2_node_frag *next = frag_next(newfrag); jffs2_add_full_dnode_to_inode() local
397 if (next) { jffs2_add_full_dnode_to_inode()
399 if (next->node) jffs2_add_full_dnode_to_inode()
400 mark_ref_normal(next->node->raw); jffs2_add_full_dnode_to_inode()
427 ret = ret->next; jffs2_get_ino_cache()
449 prev = &(*prev)->next; jffs2_add_ino_cache()
451 new->next = *prev; jffs2_add_ino_cache()
470 prev = &(*prev)->next; jffs2_del_ino_cache()
473 *prev = old->next; jffs2_del_ino_cache()
490 struct jffs2_inode_cache *this, *next; jffs2_free_ino_caches() local
495 next = this->next; jffs2_free_ino_caches()
498 this = next; jffs2_free_ino_caches()
507 struct jffs2_raw_node_ref *this, *next; jffs2_free_raw_node_refs() local
513 next = this[REFS_PER_BLOCK].next_in_ino; jffs2_free_raw_node_refs()
515 next = NULL; jffs2_free_raw_node_refs()
518 this = next; jffs2_free_raw_node_refs()
528 struct rb_node *next; jffs2_lookup_node_frag() local
534 next = fragtree->rb_node; jffs2_lookup_node_frag()
536 while(next) { jffs2_lookup_node_frag()
537 frag = rb_entry(next, struct jffs2_node_frag, rb); jffs2_lookup_node_frag()
543 next = frag->rb.rb_right; jffs2_lookup_node_frag()
545 next = frag->rb.rb_left; jffs2_lookup_node_frag()
567 struct jffs2_node_frag *frag, *next; jffs2_kill_fragtree() local
570 rbtree_postorder_for_each_entry_safe(frag, next, root, rb) { rbtree_postorder_for_each_entry_safe()
734 pr_crit("next %p (0x%08x-0x%08x)\n", __jffs2_ref_totlen()
738 pr_crit("No next ref. jeb->last_node is %p\n", __jffs2_ref_totlen()
/linux-4.4.14/net/sched/
H A Dsch_sfq.c99 sfq_index next; member in struct:sfq_head
107 sfq_index next; /* next slot in sfq RR chain */ member in struct:sfq_slot
209 n = q->dep[qlen].next; sfq_link()
211 slot->dep.next = n; sfq_link()
214 q->dep[qlen].next = x; /* sfq_dep_head(q, p)->next = x */ sfq_link()
220 n = q->slots[x].dep.next; \
222 sfq_dep_head(q, p)->next = n; \
261 skb->prev->next = (struct sk_buff *)slot; slot_dequeue_tail()
262 skb->next = skb->prev = NULL; slot_dequeue_tail()
271 slot->skblist_next = skb->next; slot_dequeue_head()
272 skb->next->prev = (struct sk_buff *)slot; slot_dequeue_head()
273 skb->next = skb->prev = NULL; slot_dequeue_head()
287 skb->next = (struct sk_buff *)slot; slot_queue_add()
288 slot->skblist_prev->next = skb; slot_queue_add()
302 x = q->dep[d].next; sfq_drop()
318 x = q->tail->next; sfq_drop()
320 q->tail->next = slot->next; sfq_drop()
368 x = q->dep[0].next; /* get a free slot */ sfq_enqueue()
447 slot->next = x; sfq_enqueue()
449 slot->next = q->tail->next; sfq_enqueue()
450 q->tail->next = x; sfq_enqueue()
489 a = q->tail->next; sfq_dequeue()
505 next_a = slot->next; sfq_dequeue()
510 q->tail->next = next_a; sfq_dequeue()
565 x = q->dep[0].next; /* get a free slot */ sfq_rehash()
589 slot->next = x; sfq_rehash()
591 slot->next = q->tail->next; sfq_rehash()
592 q->tail->next = x; sfq_rehash()
721 q->dep[i].next = i + SFQ_MAX_FLOWS; sfq_init()
/linux-4.4.14/fs/
H A Dpnode.c16 /* return the next shared peer mount of @p */ next_peer()
19 return list_entry(p->mnt_share.next, struct mount, mnt_share); next_peer()
24 return list_entry(p->mnt_slave_list.next, struct mount, mnt_slave); first_slave()
29 return list_entry(p->mnt_slave.next, struct mount, mnt_slave); next_slave()
137 * get the next mount in the propagation tree.
157 struct mount *next = next_peer(m); propagation_next() local
158 return (next == origin) ? NULL : next; propagation_next()
159 } else if (m->mnt_slave.next != &master->mnt_slave_list) propagation_next()
171 struct mount *next; next_group() local
174 next = next_peer(m); next_group()
176 if (next == origin) next_group()
178 } else if (m->mnt_slave.next != &next->mnt_slave) next_group()
180 m = next; next_group()
185 if (m->mnt_slave.next != &master->mnt_slave_list) next_group()
190 if (master->mnt_slave.next == &m->mnt_slave) next_group()
H A Dchar_dev.c32 struct char_device_struct *next; member in struct:char_device_struct
54 for (cd = chrdevs[offset]; cd; cd = cd->next) chrdev_show()
108 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) __register_chrdev_region()
135 cd->next = *cp; __register_chrdev_region()
152 for (cp = &chrdevs[i]; *cp; cp = &(*cp)->next) __unregister_chrdev_region()
159 *cp = cd->next; __unregister_chrdev_region()
178 dev_t n, next; register_chrdev_region() local
180 for (n = from; n < to; n = next) { register_chrdev_region()
181 next = MKDEV(MAJOR(n)+1, 0); register_chrdev_region()
182 if (next > to) register_chrdev_region()
183 next = to; register_chrdev_region()
185 next - n, name); register_chrdev_region()
192 for (n = from; n < to; n = next) { register_chrdev_region()
193 next = MKDEV(MAJOR(n)+1, 0); register_chrdev_region()
194 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); register_chrdev_region()
288 dev_t n, next; unregister_chrdev_region() local
290 for (n = from; n < to; n = next) { unregister_chrdev_region()
291 next = MKDEV(MAJOR(n)+1, 0); unregister_chrdev_region()
292 if (next > to) unregister_chrdev_region()
293 next = to; unregister_chrdev_region()
294 kfree(__unregister_chrdev_region(MAJOR(n), MINOR(n), next - n)); unregister_chrdev_region()
413 inode = container_of(cdev->list.next, struct inode, i_devices); cdev_purge()
/linux-4.4.14/arch/s390/mm/
H A Dgup.c108 unsigned long next; gup_pmd_range() local
118 next = pmd_addr_end(addr, end); gup_pmd_range()
138 if (!gup_huge_pmd(pmdp, pmd, addr, next, gup_pmd_range()
141 } else if (!gup_pte_range(pmdp, pmd, addr, next, gup_pmd_range()
144 } while (pmdp++, addr = next, addr != end); gup_pmd_range()
152 unsigned long next; gup_pud_range() local
162 next = pud_addr_end(addr, end); gup_pud_range()
165 if (!gup_pmd_range(pudp, pud, addr, next, write, pages, nr)) gup_pud_range()
167 } while (pudp++, addr = next, addr != end); gup_pud_range()
181 unsigned long next, flags; __get_user_pages_fast() local
203 next = pgd_addr_end(addr, end); __get_user_pages_fast()
206 if (!gup_pud_range(pgdp, pgd, addr, next, write, pages, &nr)) __get_user_pages_fast()
208 } while (pgdp++, addr = next, addr != end); __get_user_pages_fast()
/linux-4.4.14/drivers/isdn/capi/
H A Dcapilib.c12 struct capilib_msgidqueue *next; member in struct:capilib_msgidqueue
40 np->msgidpool[i].next = np->msgidfree; mq_init()
50 np->msgidfree = mq->next; mq_enqueue()
52 mq->next = NULL; mq_enqueue()
54 np->msgidlast->next = mq; mq_enqueue()
65 for (pp = &np->msgidqueue; *pp; pp = &(*pp)->next) { mq_dequeue()
68 *pp = mq->next; mq_dequeue()
71 mq->next = np->msgidfree; mq_dequeue()
/linux-4.4.14/drivers/isdn/hysdn/
H A Dhysdn_init.c95 card->next = NULL; /*end of chain */ hysdn_pci_init_one()
97 card_last->next = card; /* pointer to next card */ hysdn_pci_init_one()
129 card_root = card_root->next; hysdn_pci_remove_one()
135 if (tmp->next == card) hysdn_pci_remove_one()
136 tmp->next = card->next; hysdn_pci_remove_one()
138 tmp = tmp->next; hysdn_pci_remove_one()
/linux-4.4.14/net/6lowpan/
H A Dnhc.h19 * @_idlen: len for the next header id and mask, should be always the same.
53 * struct lowpan_nhc - hold 6lowpan next hdr compression ifnformation
56 * @name: name of the specific next header compression
57 * @nexthdr: next header value of the protocol which should be compressed.
61 * @len: the length of the next header id and mask.
62 * @setup: callback to setup fill the next header id value and mask.
124 * lowpan_nhc_add - register a next header compression to framework
131 * lowpan_nhc_del - delete a next header compression from framework
/linux-4.4.14/kernel/debug/kdb/
H A DMakefile17 /^\#/{next} \
18 /^[ \t]*$$/{next} \
/linux-4.4.14/include/asm-generic/
H A Dswitch_to.h25 #define switch_to(prev, next, last) \
27 ((last) = __switch_to((prev), (next))); \
/linux-4.4.14/arch/ia64/kernel/
H A Dunwind_i.h49 struct unw_table *next; /* must be first member! */ member in struct:unw_table
71 UNW_WHERE_GR_SAVE /* register is saved in next general register */
83 struct unw_reg_state *next; /* next (outer) element on state stack */ member in struct:unw_reg_state
88 struct unw_labeled_state *next; /* next labeled state (or NULL) */ member in struct:unw_labeled_state
110 u8 gr_save_loc; /* next general register to use for saving a register */
161 unsigned short hint; /* hint for next script to try (or -1) */
/linux-4.4.14/arch/avr32/kernel/
H A Dswitch_to.S15 /* Switch thread context from "prev" to "next", returning "last"
18 * r10 : &next->thread
/linux-4.4.14/tools/include/asm-generic/bitops/
H A Dfind.h6 * find_next_bit - find the next set bit in a memory region
11 * Returns the bit number for the next set bit
/linux-4.4.14/drivers/gpu/drm/atmel-hlcdc/
H A Datmel_hlcdc_layer.c135 dscr->next); atmel_hlcdc_layer_update_apply()
151 dscr->next); atmel_hlcdc_layer_update_apply()
368 upd->next = upd->pending ? 0 : 1; atmel_hlcdc_layer_update_start()
370 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_start()
416 upd->slots[upd->next].configs, atmel_hlcdc_layer_update_start()
429 atmel_hlcdc_layer_update_reset(layer, upd->next); atmel_hlcdc_layer_update_rollback()
430 upd->next = -1; atmel_hlcdc_layer_update_rollback()
445 if (upd->next < 0 || upd->next > 1) atmel_hlcdc_layer_update_set_fb()
454 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_set_fb()
483 if (upd->next < 0 || upd->next > 1) atmel_hlcdc_layer_update_cfg()
489 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_cfg()
502 if (upd->next < 0 || upd->next > 1) atmel_hlcdc_layer_update_commit()
505 slot = &upd->slots[upd->next]; atmel_hlcdc_layer_update_commit()
515 upd->pending = upd->next; atmel_hlcdc_layer_update_commit()
516 upd->next = -1; atmel_hlcdc_layer_update_commit()
524 upd->next = -1; atmel_hlcdc_layer_update_commit()
544 dscr->next = dma_addr + (i * sizeof(*dscr)); atmel_hlcdc_layer_dma_init()
564 dma->dscrs[0].next); atmel_hlcdc_layer_dma_cleanup()
595 upd->next = -1; atmel_hlcdc_layer_update_init()
/linux-4.4.14/block/partitions/
H A Dmsdos.c114 * table start). The second is a pointer to the next logical partition
139 if (state->next == state->limit) parse_extended()
152 * the 2nd entry is the next extended partition, or empty, parse_extended()
163 sector_t offs, size, next; parse_extended() local
172 next = this_sector + offs; parse_extended()
176 if (next < first_sector) parse_extended()
178 if (next + size > first_sector + first_size) parse_extended()
182 put_partition(state, state->next, next, size); parse_extended()
183 set_info(state, state->next, disksig); parse_extended()
185 state->parts[state->next].flags = ADDPART_FLAG_RAID; parse_extended()
187 if (++state->next == state->limit) parse_extended()
195 * It should be a link to the next logical partition. parse_extended()
248 for (i = 0; i < max_nparts && state->next < state->limit; i++) { parse_solaris_x86()
258 put_partition(state, state->next++, parse_solaris_x86()
297 if (state->next == state->limit) parse_bsd()
310 put_partition(state, state->next++, bsd_start, bsd_size); parse_bsd()
376 if (state->next == state->limit) parse_unixware()
380 put_partition(state, state->next++, parse_unixware()
420 if (state->next == state->limit) parse_minix()
424 put_partition(state, state->next++, parse_minix()
527 state->next = 5; msdos_partition()
/linux-4.4.14/drivers/pcmcia/
H A Dsa1111_generic.h7 struct sa1111_pcmcia_socket *next; member in struct:sa1111_pcmcia_socket
/linux-4.4.14/arch/um/os-Linux/
H A Dirq.c47 irq_fd = irq_fd->next; os_waiting_for_events()
106 if (*last_irq_ptr2 == &old_fd->next) os_free_irq_by_cb()
109 *prev = (*prev)->next; os_free_irq_by_cb()
115 prev = &(*prev)->next; os_free_irq_by_cb()
/linux-4.4.14/arch/unicore32/kernel/
H A Dtime.c39 unsigned long next, oscr; puv3_osmr0_set_next_event() local
42 next = readl(OST_OSCR) + delta; puv3_osmr0_set_next_event()
43 writel(next, OST_OSMR0); puv3_osmr0_set_next_event()
46 return (signed)(next - oscr) <= MIN_OSCR_DELTA ? -ETIME : 0; puv3_osmr0_set_next_event()
/linux-4.4.14/include/linux/raid/
H A Dxor.h10 struct xor_block_template *next; member in struct:xor_block_template
/linux-4.4.14/firmware/dsp56k/
H A Dbootstrap.asm68 next jclr #0,X:<<M_HSR,* label
87 _load_P jmp next
91 _load_X jmp next
95 _load_Y jmp next
/linux-4.4.14/tools/perf/util/
H A Drblist.c107 struct rb_node *pos, *next = rb_first(&rblist->entries); rblist__delete() local
109 while (next) { rblist__delete()
110 pos = next; rblist__delete()
111 next = rb_next(pos); rblist__delete()
/linux-4.4.14/drivers/media/pci/cobalt/
H A Dcobalt-omnitek.c167 dma_addr_t next = desc->bus; descriptor_list_create() local
176 WARN_ON(next & 3); descriptor_list_create()
221 next += sizeof(struct sg_dma_descriptor); descriptor_list_create()
222 d->next_h = (u32)((u64)next >> 32); descriptor_list_create()
223 d->next_l = (u32)next | descriptor_list_create()
265 next += sizeof(struct sg_dma_descriptor); descriptor_list_create()
275 d->next_h = (u32)((u64)next >> 32); descriptor_list_create()
276 d->next_l = (u32)next | (to_pci ? WRITE_TO_PCI : 0); descriptor_list_create()
284 struct sg_dma_desc_info *next) descriptor_list_chain()
289 if (next == NULL) { descriptor_list_chain()
293 d->next_h = (u32)((u64)next->bus >> 32); descriptor_list_chain()
294 d->next_l = (u32)next->bus | direction | INTERRUPT_ENABLE; descriptor_list_chain()
283 descriptor_list_chain(struct sg_dma_desc_info *this, struct sg_dma_desc_info *next) descriptor_list_chain() argument
/linux-4.4.14/fs/jfs/
H A Djfs_txnmgr.h48 lid_t next; /* index of first tlock of tid */ member in struct:tblock
96 lid_t next; /* 2: index next lockword on tid locklist member in struct:tlock
97 * next lockword on freelist
171 lid_t next; /* 2: next linelock */ member in struct:linelock
187 lid_t next; /* 2: */ member in struct:xtlock
212 * N.B. next field may be set by linelock, and should not
214 * N.B. index of the first pxdlock specifies index of next
218 lid_t next; /* 2: */ member in struct:maplock
221 u8 index; /* 2: next free maplock index */
246 lid_t next; /* 2: */ member in struct:xdlistlock

Completed in 4656 milliseconds

1234567891011>>