Searched refs:prev (Results 1 - 200 of 1017) sorted by relevance

123456

/linux-4.4.14/lib/
H A Dlist_debug.c19 * the prev/next entries already!
23 struct list_head *prev, __list_add()
26 WARN(next->prev != prev, __list_add()
27 "list_add corruption. next->prev should be " __list_add()
28 "prev (%p), but was %p. (next=%p).\n", __list_add()
29 prev, next->prev, next); __list_add()
30 WARN(prev->next != next, __list_add()
31 "list_add corruption. prev->next should be " __list_add()
32 "next (%p), but was %p. (prev=%p).\n", __list_add()
33 next, prev->next, prev); __list_add()
34 WARN(new == prev || new == next, __list_add()
35 "list_add double add: new=%p, prev=%p, next=%p.\n", __list_add()
36 new, prev, next); __list_add()
37 next->prev = new; __list_add()
39 new->prev = prev; __list_add()
40 prev->next = new; __list_add()
46 struct list_head *prev, *next; __list_del_entry() local
48 prev = entry->prev; __list_del_entry()
54 WARN(prev == LIST_POISON2, __list_del_entry()
55 "list_del corruption, %p->prev is LIST_POISON2 (%p)\n", __list_del_entry()
57 WARN(prev->next != entry, __list_del_entry()
58 "list_del corruption. prev->next should be %p, " __list_del_entry()
59 "but was %p\n", entry, prev->next) || __list_del_entry()
60 WARN(next->prev != entry, __list_del_entry()
61 "list_del corruption. next->prev should be %p, " __list_del_entry()
62 "but was %p\n", entry, next->prev)) __list_del_entry()
65 __list_del(prev, next); __list_del_entry()
79 entry->prev = LIST_POISON2; list_del()
87 struct list_head *prev, struct list_head *next) __list_add_rcu()
89 WARN(next->prev != prev, __list_add_rcu()
90 "list_add_rcu corruption. next->prev should be prev (%p), but was %p. (next=%p).\n", __list_add_rcu()
91 prev, next->prev, next); __list_add_rcu()
92 WARN(prev->next != next, __list_add_rcu()
93 "list_add_rcu corruption. prev->next should be next (%p), but was %p. (prev=%p).\n", __list_add_rcu()
94 next, prev->next, prev); __list_add_rcu()
96 new->prev = prev; __list_add_rcu()
97 rcu_assign_pointer(list_next_rcu(prev), new); __list_add_rcu()
98 next->prev = new; __list_add_rcu()
22 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
86 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
H A Dplist.c36 WARN(n->prev != p || p->next != n, plist_check_prev_next()
38 "prev: %p, n: %p, p: %p\n" plist_check_prev_next()
40 t, t->next, t->prev, plist_check_prev_next()
41 p, p->next, p->prev, plist_check_prev_next()
42 n, n->next, n->prev); plist_check_prev_next()
47 struct list_head *prev = top, *next = top->next; plist_check_list() local
49 plist_check_prev_next(top, prev, next); plist_check_list()
51 prev = next; plist_check_list()
52 next = prev->next; plist_check_list()
53 plist_check_prev_next(top, prev, next); plist_check_list()
76 struct plist_node *first, *iter, *prev = NULL; plist_add() local
94 prev = iter; plist_add()
99 if (!prev || prev->prio != node->prio) plist_add()
H A Dklist.c327 * klist_prev - Ante up prev node in list.
331 * node, if there was one. Grab the prev node, increment its reference
332 * count, drop the lock, and return that prev node.
338 struct klist_node *prev; klist_prev() local
343 prev = to_klist_node(last->n_node.prev); klist_prev()
347 prev = to_klist_node(i->i_klist->k_list.prev); klist_prev()
350 while (prev != to_klist_node(&i->i_klist->k_list)) { klist_prev()
351 if (likely(!knode_dead(prev))) { klist_prev()
352 kref_get(&prev->n_ref); klist_prev()
353 i->i_cur = prev; klist_prev()
356 prev = to_klist_node(prev->n_node.prev); klist_prev()
H A Dlist_sort.c17 * sentinel head node, "prev" links not maintained.
45 * prev-link restoration pass, or maintaining the prev links
61 a->prev = tail; merge_and_restore_back_links()
65 b->prev = tail; merge_and_restore_back_links()
82 tail->next->prev = tail; merge_and_restore_back_links()
87 head->prev = tail; merge_and_restore_back_links()
119 head->prev->next = NULL; list_sort()
249 if (cur->next->prev != cur) { list_sort_test()
274 if (head.prev != cur) { list_sort_test()
/linux-4.4.14/arch/score/include/asm/
H A Dswitch_to.h6 #define switch_to(prev, next, last) \
8 (last) = resume(prev, next, task_thread_info(next)); \
H A Dmmu_context.h72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
102 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
/linux-4.4.14/arch/parisc/include/asm/
H A Dswitch_to.h8 #define switch_to(prev, next, last) do { \
9 (last) = _switch_to(prev, next); \
H A Dmmu_context.h52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument
55 if (prev != next) { switch_mm()
63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
78 switch_mm(prev,next,current); activate_mm()
/linux-4.4.14/arch/s390/include/asm/
H A Dswitch_to.h31 #define switch_to(prev,next,last) do { \
32 if (prev->mm) { \
34 save_access_regs(&prev->thread.acrs[0]); \
35 save_ri_cb(prev->thread.ri_cb); \
41 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \
43 prev = __switch_to(prev,next); \
H A Dmmu_context.h85 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
91 if (prev == next) switch_mm()
99 atomic_dec(&prev->context.attach_count); switch_mm()
101 cpumask_clear_cpu(cpu, &prev->context.cpu_attach_mask); switch_mm()
127 static inline void activate_mm(struct mm_struct *prev, activate_mm() argument
130 switch_mm(prev, next, current); activate_mm()
/linux-4.4.14/tools/firewire/
H A Dlist.h2 struct list *next, *prev; member in struct:list
9 list->prev = list; list_init()
21 new_link->prev = link->prev; list_insert()
23 new_link->prev->next = new_link; list_insert()
24 new_link->next->prev = new_link; list_insert()
42 link->prev->next = link->next; list_remove()
43 link->next->prev = link->prev; list_remove()
53 list_entry((list)->prev, type, member)
H A Dnosy-dump.c498 struct subaction *sa, *prev; handle_request_packet() local
507 prev = list_tail(&t->request_list, handle_request_packet()
510 if (!ACK_BUSY(prev->ack)) { handle_request_packet()
519 if (prev->packet.common.tcode != sa->packet.common.tcode || handle_request_packet()
520 prev->packet.common.tlabel != sa->packet.common.tlabel) { handle_request_packet()
563 struct subaction *sa, *prev; handle_response_packet() local
576 prev = list_tail(&t->response_list, struct subaction, link); handle_response_packet()
578 if (!ACK_BUSY(prev->ack)) { handle_response_packet()
585 if (prev->packet.common.tcode != sa->packet.common.tcode || handle_response_packet()
586 prev->packet.common.tlabel != sa->packet.common.tlabel) { handle_response_packet()
591 prev = list_tail(&t->request_list, struct subaction, link); handle_response_packet()
592 if (prev->ack != ACK_PENDING) { handle_response_packet()
599 if (packet_info[prev->packet.common.tcode].response_tcode != handle_response_packet()
/linux-4.4.14/include/asm-generic/
H A Dcmpxchg-local.h17 unsigned long flags, prev; __cmpxchg_local_generic() local
27 case 1: prev = *(u8 *)ptr; __cmpxchg_local_generic()
28 if (prev == old) __cmpxchg_local_generic()
31 case 2: prev = *(u16 *)ptr; __cmpxchg_local_generic()
32 if (prev == old) __cmpxchg_local_generic()
35 case 4: prev = *(u32 *)ptr; __cmpxchg_local_generic()
36 if (prev == old) __cmpxchg_local_generic()
39 case 8: prev = *(u64 *)ptr; __cmpxchg_local_generic()
40 if (prev == old) __cmpxchg_local_generic()
47 return prev; __cmpxchg_local_generic()
56 u64 prev; __cmpxchg64_local_generic() local
60 prev = *(u64 *)ptr; __cmpxchg64_local_generic()
61 if (prev == old) __cmpxchg64_local_generic()
64 return prev; __cmpxchg64_local_generic()
H A Dmutex-xchg.h94 int prev = atomic_xchg_acquire(count, 0); __mutex_fastpath_trylock() local
96 if (unlikely(prev < 0)) { __mutex_fastpath_trylock()
99 * state. If while doing so we get back a prev value of 1 __mutex_fastpath_trylock()
108 prev = atomic_xchg_acquire(count, prev); __mutex_fastpath_trylock()
109 if (prev < 0) __mutex_fastpath_trylock()
110 prev = 0; __mutex_fastpath_trylock()
113 return prev; __mutex_fastpath_trylock()
H A Dswitch_to.h25 #define switch_to(prev, next, last) \
27 ((last) = __switch_to((prev), (next))); \
H A Dmmu_context.h34 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument
/linux-4.4.14/arch/unicore32/include/asm/
H A Dswitch_to.h17 * switch_to(prev, next) should switch from task `prev' to `next'
18 * `prev' will never be the same as `next'. schedule() itself
24 #define switch_to(prev, next, last) \
26 last = __switch_to(prev, task_thread_info(prev), \
H A Dmmu_context.h49 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
54 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) switch_mm()
59 #define activate_mm(prev, next) switch_mm(prev, next, NULL)
/linux-4.4.14/arch/blackfin/include/asm/
H A Dswitch_to.h21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next);
24 #define switch_to(prev,next,last) \
26 memcpy (&task_thread_info(prev)->l1_task_info, L1_SCRATCH_TASK_INFO, \
30 (last) = resume (prev, next); \
33 #define switch_to(prev, next, last) \
35 (last) = resume(prev, next); \
H A Dmmu_context.h71 #define activate_mm(prev, next) switch_mm(prev, next, NULL)
111 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
116 __switch_mm(prev, next, tsk); switch_mm()
160 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
163 __switch_mm(prev, next, tsk); switch_mm()
/linux-4.4.14/drivers/media/platform/omap3isp/
H A Disppreview.c152 preview_config_luma_enhancement(struct isp_prev_device *prev, preview_config_luma_enhancement() argument
155 struct isp_device *isp = to_isp_device(prev); preview_config_luma_enhancement()
171 preview_enable_luma_enhancement(struct isp_prev_device *prev, bool enable) preview_enable_luma_enhancement() argument
173 struct isp_device *isp = to_isp_device(prev); preview_enable_luma_enhancement()
186 static void preview_enable_invalaw(struct isp_prev_device *prev, bool enable) preview_enable_invalaw() argument
188 struct isp_device *isp = to_isp_device(prev); preview_enable_invalaw()
201 static void preview_config_hmed(struct isp_prev_device *prev, preview_config_hmed() argument
204 struct isp_device *isp = to_isp_device(prev); preview_config_hmed()
216 static void preview_enable_hmed(struct isp_prev_device *prev, bool enable) preview_enable_hmed() argument
218 struct isp_device *isp = to_isp_device(prev); preview_enable_hmed()
236 static void preview_config_cfa(struct isp_prev_device *prev, preview_config_cfa() argument
245 const unsigned int *order = cfa_coef_order[prev->params.cfa_order]; preview_config_cfa()
247 struct isp_device *isp = to_isp_device(prev); preview_config_cfa()
272 preview_config_chroma_suppression(struct isp_prev_device *prev, preview_config_chroma_suppression() argument
275 struct isp_device *isp = to_isp_device(prev); preview_config_chroma_suppression()
288 preview_enable_chroma_suppression(struct isp_prev_device *prev, bool enable) preview_enable_chroma_suppression() argument
290 struct isp_device *isp = to_isp_device(prev); preview_enable_chroma_suppression()
306 preview_config_whitebalance(struct isp_prev_device *prev, preview_config_whitebalance() argument
309 struct isp_device *isp = to_isp_device(prev); preview_config_whitebalance()
345 preview_config_blkadj(struct isp_prev_device *prev, preview_config_blkadj() argument
348 struct isp_device *isp = to_isp_device(prev); preview_config_blkadj()
361 preview_config_rgb_blending(struct isp_prev_device *prev, preview_config_rgb_blending() argument
364 struct isp_device *isp = to_isp_device(prev); preview_config_rgb_blending()
399 preview_config_csc(struct isp_prev_device *prev, preview_config_csc() argument
402 struct isp_device *isp = to_isp_device(prev); preview_config_csc()
431 preview_config_yc_range(struct isp_prev_device *prev, preview_config_yc_range() argument
434 struct isp_device *isp = to_isp_device(prev); preview_config_yc_range()
449 preview_config_dcor(struct isp_prev_device *prev, preview_config_dcor() argument
452 struct isp_device *isp = to_isp_device(prev); preview_config_dcor()
471 static void preview_enable_dcor(struct isp_prev_device *prev, bool enable) preview_enable_dcor() argument
473 struct isp_device *isp = to_isp_device(prev); preview_enable_dcor()
487 preview_enable_drkframe_capture(struct isp_prev_device *prev, bool enable) preview_enable_drkframe_capture() argument
489 struct isp_device *isp = to_isp_device(prev); preview_enable_drkframe_capture()
502 static void preview_enable_drkframe(struct isp_prev_device *prev, bool enable) preview_enable_drkframe() argument
504 struct isp_device *isp = to_isp_device(prev); preview_enable_drkframe()
518 preview_config_noisefilter(struct isp_prev_device *prev, preview_config_noisefilter() argument
521 struct isp_device *isp = to_isp_device(prev); preview_config_noisefilter()
538 preview_enable_noisefilter(struct isp_prev_device *prev, bool enable) preview_enable_noisefilter() argument
540 struct isp_device *isp = to_isp_device(prev); preview_enable_noisefilter()
554 preview_config_gammacorrn(struct isp_prev_device *prev, preview_config_gammacorrn() argument
557 struct isp_device *isp = to_isp_device(prev); preview_config_gammacorrn()
587 preview_enable_gammacorrn(struct isp_prev_device *prev, bool enable) preview_enable_gammacorrn() argument
589 struct isp_device *isp = to_isp_device(prev); preview_enable_gammacorrn()
605 preview_config_contrast(struct isp_prev_device *prev, preview_config_contrast() argument
608 struct isp_device *isp = to_isp_device(prev); preview_config_contrast()
619 preview_config_brightness(struct isp_prev_device *prev, preview_config_brightness() argument
622 struct isp_device *isp = to_isp_device(prev); preview_config_brightness()
636 preview_update_contrast(struct isp_prev_device *prev, u8 contrast) preview_update_contrast() argument
641 spin_lock_irqsave(&prev->params.lock, flags); preview_update_contrast()
642 params = (prev->params.active & OMAP3ISP_PREV_CONTRAST) preview_update_contrast()
643 ? &prev->params.params[0] : &prev->params.params[1]; preview_update_contrast()
649 spin_unlock_irqrestore(&prev->params.lock, flags); preview_update_contrast()
658 preview_update_brightness(struct isp_prev_device *prev, u8 brightness) preview_update_brightness() argument
663 spin_lock_irqsave(&prev->params.lock, flags); preview_update_brightness()
664 params = (prev->params.active & OMAP3ISP_PREV_BRIGHTNESS) preview_update_brightness()
665 ? &prev->params.params[0] : &prev->params.params[1]; preview_update_brightness()
671 spin_unlock_irqrestore(&prev->params.lock, flags); preview_update_brightness()
675 preview_params_lock(struct isp_prev_device *prev, u32 update, bool shadow) preview_params_lock() argument
677 u32 active = prev->params.active; preview_params_lock()
681 prev->params.params[0].busy |= ~active & update; preview_params_lock()
682 prev->params.params[1].busy |= active & update; preview_params_lock()
685 update = (prev->params.params[0].update & active) preview_params_lock()
686 | (prev->params.params[1].update & ~active); preview_params_lock()
688 prev->params.params[0].busy |= active & update; preview_params_lock()
689 prev->params.params[1].busy |= ~active & update; preview_params_lock()
696 preview_params_unlock(struct isp_prev_device *prev, u32 update, bool shadow) preview_params_unlock() argument
698 u32 active = prev->params.active; preview_params_unlock()
704 prev->params.params[0].update |= (~active & update); preview_params_unlock()
705 prev->params.params[1].update |= (active & update); preview_params_unlock()
706 prev->params.params[0].busy &= active; preview_params_unlock()
707 prev->params.params[1].busy &= ~active; preview_params_unlock()
712 prev->params.params[0].update &= ~(active & update); preview_params_unlock()
713 prev->params.params[1].update &= ~(~active & update); preview_params_unlock()
714 prev->params.params[0].busy &= ~active; preview_params_unlock()
715 prev->params.params[1].busy &= active; preview_params_unlock()
719 static void preview_params_switch(struct isp_prev_device *prev) preview_params_switch() argument
727 to_switch = (prev->params.params[0].update & ~prev->params.active) preview_params_switch()
728 | (prev->params.params[1].update & prev->params.active); preview_params_switch()
729 to_switch &= ~(prev->params.params[0].busy | preview_params_switch()
730 prev->params.params[1].busy); preview_params_switch()
734 prev->params.active ^= to_switch; preview_params_switch()
739 prev->params.params[0].update &= ~(~prev->params.active & to_switch); preview_params_switch()
740 prev->params.params[1].update &= ~(prev->params.active & to_switch); preview_params_switch()
856 * @prev: ISP preview engine
862 static int preview_config(struct isp_prev_device *prev, preview_config() argument
875 spin_lock_irqsave(&prev->params.lock, flags); preview_config()
876 preview_params_lock(prev, cfg->update, true); preview_config()
877 active = prev->params.active; preview_config()
878 spin_unlock_irqrestore(&prev->params.lock, flags); preview_config()
890 params = &prev->params.params[!!(active & bit)]; preview_config()
912 spin_lock_irqsave(&prev->params.lock, flags); preview_config()
913 preview_params_unlock(prev, update, true); preview_config()
914 preview_params_switch(prev); preview_config()
915 spin_unlock_irqrestore(&prev->params.lock, flags); preview_config()
922 * @prev: pointer to preview private structure
928 static void preview_setup_hw(struct isp_prev_device *prev, u32 update, preview_setup_hw() argument
944 params = &prev->params.params[!(active & bit)]; preview_setup_hw()
948 attr->config(prev, params); preview_setup_hw()
950 attr->enable(prev, true); preview_setup_hw()
953 attr->enable(prev, false); preview_setup_hw()
960 * @prev: pointer to previewer private structure
963 static void preview_config_ycpos(struct isp_prev_device *prev, u32 pixelcode) preview_config_ycpos() argument
965 struct isp_device *isp = to_isp_device(prev); preview_config_ycpos()
988 static void preview_config_averager(struct isp_prev_device *prev, u8 average) preview_config_averager() argument
990 struct isp_device *isp = to_isp_device(prev); preview_config_averager()
1000 * @prev: The preview engine
1011 static void preview_config_input_format(struct isp_prev_device *prev, preview_config_input_format() argument
1014 struct isp_device *isp = to_isp_device(prev); preview_config_input_format()
1026 prev->params.cfa_order = 0; preview_config_input_format()
1029 prev->params.cfa_order = 1; preview_config_input_format()
1032 prev->params.cfa_order = 2; preview_config_input_format()
1035 prev->params.cfa_order = 3; preview_config_input_format()
1048 params = (prev->params.active & OMAP3ISP_PREV_CFA) preview_config_input_format()
1049 ? &prev->params.params[0] : &prev->params.params[1]; preview_config_input_format()
1051 preview_config_cfa(prev, params); preview_config_input_format()
1065 static void preview_config_input_size(struct isp_prev_device *prev, u32 active) preview_config_input_size() argument
1067 const struct v4l2_mbus_framefmt *format = &prev->formats[PREV_PAD_SINK]; preview_config_input_size()
1068 struct isp_device *isp = to_isp_device(prev); preview_config_input_size()
1069 unsigned int sph = prev->crop.left; preview_config_input_size()
1070 unsigned int eph = prev->crop.left + prev->crop.width - 1; preview_config_input_size()
1071 unsigned int slv = prev->crop.top; preview_config_input_size()
1072 unsigned int elv = prev->crop.top + prev->crop.height - 1; preview_config_input_size()
1083 features = (prev->params.params[0].features & active) preview_config_input_size()
1084 | (prev->params.params[1].features & ~active); preview_config_input_size()
1107 * @prev: Preview module
1116 preview_config_inlineoffset(struct isp_prev_device *prev, u32 offset) preview_config_inlineoffset() argument
1118 struct isp_device *isp = to_isp_device(prev); preview_config_inlineoffset()
1130 static void preview_set_inaddr(struct isp_prev_device *prev, u32 addr) preview_set_inaddr() argument
1132 struct isp_device *isp = to_isp_device(prev); preview_set_inaddr()
1143 static void preview_config_outlineoffset(struct isp_prev_device *prev, preview_config_outlineoffset() argument
1146 struct isp_device *isp = to_isp_device(prev); preview_config_outlineoffset()
1158 static void preview_set_outaddr(struct isp_prev_device *prev, u32 addr) preview_set_outaddr() argument
1160 struct isp_device *isp = to_isp_device(prev); preview_set_outaddr()
1165 static void preview_adjust_bandwidth(struct isp_prev_device *prev) preview_adjust_bandwidth() argument
1167 struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); preview_adjust_bandwidth()
1168 struct isp_device *isp = to_isp_device(prev); preview_adjust_bandwidth()
1169 const struct v4l2_mbus_framefmt *ifmt = &prev->formats[PREV_PAD_SINK]; preview_adjust_bandwidth()
1179 if (prev->input != PREVIEW_INPUT_MEMORY) { preview_adjust_bandwidth()
1218 int omap3isp_preview_busy(struct isp_prev_device *prev) omap3isp_preview_busy() argument
1220 struct isp_device *isp = to_isp_device(prev); omap3isp_preview_busy()
1231 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_restore_context() local
1234 prev->params.params[0].update = prev->params.active & update; omap3isp_preview_restore_context()
1235 prev->params.params[1].update = ~prev->params.active & update; omap3isp_preview_restore_context()
1237 preview_setup_hw(prev, update, prev->params.active); omap3isp_preview_restore_context()
1239 prev->params.params[0].update = 0; omap3isp_preview_restore_context()
1240 prev->params.params[1].update = 0; omap3isp_preview_restore_context()
1250 static void preview_print_status(struct isp_prev_device *prev) preview_print_status() argument
1252 struct isp_device *isp = to_isp_device(prev); preview_print_status()
1298 * @prev: pointer to previewer private structure
1300 static void preview_init_params(struct isp_prev_device *prev) preview_init_params() argument
1305 spin_lock_init(&prev->params.lock); preview_init_params()
1307 prev->params.active = ~0; preview_init_params()
1308 prev->params.params[0].busy = 0; preview_init_params()
1309 prev->params.params[0].update = OMAP3ISP_PREV_FEATURES_END - 1; preview_init_params()
1310 prev->params.params[1].busy = 0; preview_init_params()
1311 prev->params.params[1].update = 0; preview_init_params()
1313 params = &prev->params.params[0]; preview_init_params()
1361 * @prev: pointer to previewer private structure
1364 static unsigned int preview_max_out_width(struct isp_prev_device *prev) preview_max_out_width() argument
1366 struct isp_device *isp = to_isp_device(prev); preview_max_out_width()
1381 static void preview_configure(struct isp_prev_device *prev) preview_configure() argument
1383 struct isp_device *isp = to_isp_device(prev); preview_configure()
1390 spin_lock_irqsave(&prev->params.lock, flags); preview_configure()
1392 update = preview_params_lock(prev, 0, false); preview_configure()
1393 active = prev->params.active; preview_configure()
1394 spin_unlock_irqrestore(&prev->params.lock, flags); preview_configure()
1397 format = &prev->formats[PREV_PAD_SINK]; preview_configure()
1400 preview_adjust_bandwidth(prev); preview_configure()
1402 preview_config_input_format(prev, info); preview_configure()
1403 preview_config_input_size(prev, active); preview_configure()
1405 if (prev->input == PREVIEW_INPUT_CCDC) preview_configure()
1406 preview_config_inlineoffset(prev, 0); preview_configure()
1408 preview_config_inlineoffset(prev, ALIGN(format->width, 0x20) * preview_configure()
1411 preview_setup_hw(prev, update, active); preview_configure()
1414 format = &prev->formats[PREV_PAD_SOURCE]; preview_configure()
1416 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_configure()
1423 if (prev->output & PREVIEW_OUTPUT_RESIZER) preview_configure()
1430 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_configure()
1431 preview_config_outlineoffset(prev, preview_configure()
1434 preview_config_averager(prev, 0); preview_configure()
1435 preview_config_ycpos(prev, format->code); preview_configure()
1437 spin_lock_irqsave(&prev->params.lock, flags); preview_configure()
1438 preview_params_unlock(prev, update, false); preview_configure()
1439 spin_unlock_irqrestore(&prev->params.lock, flags); preview_configure()
1446 static void preview_enable_oneshot(struct isp_prev_device *prev) preview_enable_oneshot() argument
1448 struct isp_device *isp = to_isp_device(prev); preview_enable_oneshot()
1454 if (prev->input == PREVIEW_INPUT_MEMORY) preview_enable_oneshot()
1462 void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev) omap3isp_preview_isr_frame_sync() argument
1470 if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS && omap3isp_preview_isr_frame_sync()
1471 prev->video_out.dmaqueue_flags & ISP_VIDEO_DMAQUEUE_QUEUED) { omap3isp_preview_isr_frame_sync()
1472 preview_enable_oneshot(prev); omap3isp_preview_isr_frame_sync()
1473 isp_video_dmaqueue_flags_clr(&prev->video_out); omap3isp_preview_isr_frame_sync()
1477 static void preview_isr_buffer(struct isp_prev_device *prev) preview_isr_buffer() argument
1479 struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); preview_isr_buffer()
1483 if (prev->input == PREVIEW_INPUT_MEMORY) { preview_isr_buffer()
1484 buffer = omap3isp_video_buffer_next(&prev->video_in); preview_isr_buffer()
1486 preview_set_inaddr(prev, buffer->dma); preview_isr_buffer()
1490 if (prev->output & PREVIEW_OUTPUT_MEMORY) { preview_isr_buffer()
1491 buffer = omap3isp_video_buffer_next(&prev->video_out); preview_isr_buffer()
1493 preview_set_outaddr(prev, buffer->dma); preview_isr_buffer()
1499 switch (prev->state) { preview_isr_buffer()
1511 preview_enable_oneshot(prev); preview_isr_buffer()
1525 void omap3isp_preview_isr(struct isp_prev_device *prev) omap3isp_preview_isr() argument
1531 if (omap3isp_module_sync_is_stopping(&prev->wait, &prev->stopping)) omap3isp_preview_isr()
1534 spin_lock_irqsave(&prev->params.lock, flags); omap3isp_preview_isr()
1535 preview_params_switch(prev); omap3isp_preview_isr()
1536 update = preview_params_lock(prev, 0, false); omap3isp_preview_isr()
1537 active = prev->params.active; omap3isp_preview_isr()
1538 spin_unlock_irqrestore(&prev->params.lock, flags); omap3isp_preview_isr()
1540 preview_setup_hw(prev, update, active); omap3isp_preview_isr()
1541 preview_config_input_size(prev, active); omap3isp_preview_isr()
1543 if (prev->input == PREVIEW_INPUT_MEMORY || omap3isp_preview_isr()
1544 prev->output & PREVIEW_OUTPUT_MEMORY) omap3isp_preview_isr()
1545 preview_isr_buffer(prev); omap3isp_preview_isr()
1546 else if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS) omap3isp_preview_isr()
1547 preview_enable_oneshot(prev); omap3isp_preview_isr()
1549 spin_lock_irqsave(&prev->params.lock, flags); omap3isp_preview_isr()
1550 preview_params_unlock(prev, update, false); omap3isp_preview_isr()
1551 spin_unlock_irqrestore(&prev->params.lock, flags); omap3isp_preview_isr()
1561 struct isp_prev_device *prev = &video->isp->isp_prev; preview_video_queue() local
1564 preview_set_inaddr(prev, buffer->dma); preview_video_queue()
1567 preview_set_outaddr(prev, buffer->dma); preview_video_queue()
1586 struct isp_prev_device *prev = preview_s_ctrl() local
1591 preview_update_brightness(prev, ctrl->val); preview_s_ctrl()
1594 preview_update_contrast(prev, ctrl->val); preview_s_ctrl()
1614 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_ioctl() local
1618 return preview_config(prev, arg); preview_ioctl()
1633 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_stream() local
1634 struct isp_video *video_out = &prev->video_out; preview_set_stream()
1635 struct isp_device *isp = to_isp_device(prev); preview_set_stream()
1636 struct device *dev = to_device(prev); preview_set_stream()
1638 if (prev->state == ISP_PIPELINE_STREAM_STOPPED) { preview_set_stream()
1643 preview_configure(prev); preview_set_stream()
1644 atomic_set(&prev->stopping, 0); preview_set_stream()
1645 preview_print_status(prev); preview_set_stream()
1650 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_set_stream()
1654 !(prev->output & PREVIEW_OUTPUT_MEMORY)) preview_set_stream()
1655 preview_enable_oneshot(prev); preview_set_stream()
1661 if (prev->input == PREVIEW_INPUT_MEMORY) preview_set_stream()
1663 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_set_stream()
1666 preview_enable_oneshot(prev); preview_set_stream()
1670 if (omap3isp_module_sync_idle(&sd->entity, &prev->wait, preview_set_stream()
1671 &prev->stopping)) preview_set_stream()
1680 prev->state = enable; preview_set_stream()
1685 __preview_get_format(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, __preview_get_format() argument
1689 return v4l2_subdev_get_try_format(&prev->subdev, cfg, pad); __preview_get_format()
1691 return &prev->formats[pad]; __preview_get_format()
1695 __preview_get_crop(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, __preview_get_crop() argument
1699 return v4l2_subdev_get_try_crop(&prev->subdev, cfg, PREV_PAD_SINK); __preview_get_crop()
1701 return &prev->crop; __preview_get_crop()
1725 * @prev: ISP preview engine
1734 static void preview_try_format(struct isp_prev_device *prev, preview_try_format() argument
1754 if (prev->input == PREVIEW_INPUT_MEMORY) { preview_try_format()
1756 preview_max_out_width(prev)); preview_try_format()
1776 *fmt = *__preview_get_format(prev, cfg, PREV_PAD_SINK, which); preview_try_format()
1794 crop = __preview_get_crop(prev, cfg, which); preview_try_format()
1807 * @prev: ISP preview engine
1817 static void preview_try_crop(struct isp_prev_device *prev, preview_try_crop() argument
1831 if (prev->input == PREVIEW_INPUT_CCDC) { preview_try_crop()
1895 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_enum_frame_size() local
1904 preview_try_format(prev, cfg, fse->pad, &format, fse->which); preview_enum_frame_size()
1914 preview_try_format(prev, cfg, fse->pad, &format, fse->which); preview_enum_frame_size()
1935 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_get_selection() local
1948 format = __preview_get_format(prev, cfg, PREV_PAD_SINK, preview_get_selection()
1950 preview_try_crop(prev, format, &sel->r); preview_get_selection()
1954 sel->r = *__preview_get_crop(prev, cfg, sel->which); preview_get_selection()
1978 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_selection() local
1986 if (prev->state != ISP_PIPELINE_STREAM_STOPPED) preview_set_selection()
1994 sel->r = *__preview_get_crop(prev, cfg, sel->which); preview_set_selection()
1998 format = __preview_get_format(prev, cfg, PREV_PAD_SINK, sel->which); preview_set_selection()
1999 preview_try_crop(prev, format, &sel->r); preview_set_selection()
2000 *__preview_get_crop(prev, cfg, sel->which) = sel->r; preview_set_selection()
2003 format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, sel->which); preview_set_selection()
2004 preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, sel->which); preview_set_selection()
2019 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_get_format() local
2022 format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); preview_get_format()
2040 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_format() local
2044 format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); preview_set_format()
2048 preview_try_format(prev, cfg, fmt->pad, &fmt->format, fmt->which); preview_set_format()
2054 crop = __preview_get_crop(prev, cfg, fmt->which); preview_set_format()
2060 preview_try_crop(prev, &fmt->format, crop); preview_set_format()
2063 format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, preview_set_format()
2065 preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, preview_set_format()
2146 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_link_setup() local
2152 if (prev->input == PREVIEW_INPUT_CCDC) preview_link_setup()
2154 prev->input = PREVIEW_INPUT_MEMORY; preview_link_setup()
2156 if (prev->input == PREVIEW_INPUT_MEMORY) preview_link_setup()
2157 prev->input = PREVIEW_INPUT_NONE; preview_link_setup()
2164 if (prev->input == PREVIEW_INPUT_MEMORY) preview_link_setup()
2166 prev->input = PREVIEW_INPUT_CCDC; preview_link_setup()
2168 if (prev->input == PREVIEW_INPUT_CCDC) preview_link_setup()
2169 prev->input = PREVIEW_INPUT_NONE; preview_link_setup()
2181 if (prev->output & ~PREVIEW_OUTPUT_MEMORY) preview_link_setup()
2183 prev->output |= PREVIEW_OUTPUT_MEMORY; preview_link_setup()
2185 prev->output &= ~PREVIEW_OUTPUT_MEMORY; preview_link_setup()
2192 if (prev->output & ~PREVIEW_OUTPUT_RESIZER) preview_link_setup()
2194 prev->output |= PREVIEW_OUTPUT_RESIZER; preview_link_setup()
2196 prev->output &= ~PREVIEW_OUTPUT_RESIZER; preview_link_setup()
2213 void omap3isp_preview_unregister_entities(struct isp_prev_device *prev) omap3isp_preview_unregister_entities() argument
2215 v4l2_device_unregister_subdev(&prev->subdev); omap3isp_preview_unregister_entities()
2216 omap3isp_video_unregister(&prev->video_in); omap3isp_preview_unregister_entities()
2217 omap3isp_video_unregister(&prev->video_out); omap3isp_preview_unregister_entities()
2220 int omap3isp_preview_register_entities(struct isp_prev_device *prev, omap3isp_preview_register_entities() argument
2226 ret = v4l2_device_register_subdev(vdev, &prev->subdev); omap3isp_preview_register_entities()
2230 ret = omap3isp_video_register(&prev->video_in, vdev); omap3isp_preview_register_entities()
2234 ret = omap3isp_video_register(&prev->video_out, vdev); omap3isp_preview_register_entities()
2241 omap3isp_preview_unregister_entities(prev); omap3isp_preview_register_entities()
2251 * @prev : Pointer to preview structure
2254 static int preview_init_entities(struct isp_prev_device *prev) preview_init_entities() argument
2256 struct v4l2_subdev *sd = &prev->subdev; preview_init_entities()
2257 struct media_pad *pads = prev->pads; preview_init_entities()
2261 prev->input = PREVIEW_INPUT_NONE; preview_init_entities()
2267 v4l2_set_subdevdata(sd, prev); preview_init_entities()
2270 v4l2_ctrl_handler_init(&prev->ctrls, 2); preview_init_entities()
2271 v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_BRIGHTNESS, preview_init_entities()
2274 v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_CONTRAST, preview_init_entities()
2277 v4l2_ctrl_handler_setup(&prev->ctrls); preview_init_entities()
2278 sd->ctrl_handler = &prev->ctrls; preview_init_entities()
2295 prev->video_in.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; preview_init_entities()
2296 prev->video_in.ops = &preview_video_ops; preview_init_entities()
2297 prev->video_in.isp = to_isp_device(prev); preview_init_entities()
2298 prev->video_in.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; preview_init_entities()
2299 prev->video_in.bpl_alignment = 64; preview_init_entities()
2300 prev->video_out.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; preview_init_entities()
2301 prev->video_out.ops = &preview_video_ops; preview_init_entities()
2302 prev->video_out.isp = to_isp_device(prev); preview_init_entities()
2303 prev->video_out.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; preview_init_entities()
2304 prev->video_out.bpl_alignment = 32; preview_init_entities()
2306 ret = omap3isp_video_init(&prev->video_in, "preview"); preview_init_entities()
2310 ret = omap3isp_video_init(&prev->video_out, "preview"); preview_init_entities()
2315 ret = media_entity_create_link(&prev->video_in.video.entity, 0, preview_init_entities()
2316 &prev->subdev.entity, PREV_PAD_SINK, 0); preview_init_entities()
2320 ret = media_entity_create_link(&prev->subdev.entity, PREV_PAD_SOURCE, preview_init_entities()
2321 &prev->video_out.video.entity, 0, 0); preview_init_entities()
2328 omap3isp_video_cleanup(&prev->video_out); preview_init_entities()
2330 omap3isp_video_cleanup(&prev->video_in); preview_init_entities()
2332 media_entity_cleanup(&prev->subdev.entity); preview_init_entities()
2343 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_init() local
2345 init_waitqueue_head(&prev->wait); omap3isp_preview_init()
2347 preview_init_params(prev); omap3isp_preview_init()
2349 return preview_init_entities(prev); omap3isp_preview_init()
2354 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_cleanup() local
2356 v4l2_ctrl_handler_free(&prev->ctrls); omap3isp_preview_cleanup()
2357 omap3isp_video_cleanup(&prev->video_in); omap3isp_preview_cleanup()
2358 omap3isp_video_cleanup(&prev->video_out); omap3isp_preview_cleanup()
2359 media_entity_cleanup(&prev->subdev.entity); omap3isp_preview_cleanup()
/linux-4.4.14/arch/frv/include/asm/
H A Dswitch_to.h18 * switch_to(prev, next) should switch from task `prev' to `next'
19 * `prev' will never be the same as `next'.
25 struct task_struct *prev);
27 #define switch_to(prev, next, last) \
29 (prev)->thread.sched_lr = \
31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
H A Dmmu_context.h35 #define switch_mm(prev, next, tsk) \
37 if (prev != next) \
38 change_mm_context(&prev->context, &next->context, next->pgd); \
41 #define activate_mm(prev, next) \
43 change_mm_context(&prev->context, &next->context, next->pgd); \
/linux-4.4.14/arch/cris/include/asm/
H A Dswitch_to.h8 extern struct task_struct *resume(struct task_struct *prev, struct task_struct *next, int);
9 #define switch_to(prev,next,last) last = resume(prev,next, \
H A Dmmu_context.h9 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
14 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
19 switch_mm(prev, next, NULL); activate_mm()
H A Dfasttimer.h21 struct fast_timer *prev; member in struct:fast_timer
/linux-4.4.14/arch/mn10300/include/asm/
H A Dswitch_to.h22 #define switch_fpu(prev, next) \
24 if ((prev)->thread.fpu_flags & THREAD_HAS_FPU) { \
25 (prev)->thread.fpu_flags &= ~THREAD_HAS_FPU; \
26 (prev)->thread.uregs->epsw &= ~EPSW_FE; \
27 fpu_save(&(prev)->thread.fpu_state); \
31 #define switch_fpu(prev, next) do {} while (0)
36 struct task_struct *__switch_to(struct thread_struct *prev,
40 #define switch_to(prev, next, last) \
42 switch_fpu(prev, next); \
44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
H A Dmmu_context.h139 * @prev: The outgoing MM context.
143 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
148 if (prev != next) { switch_mm()
159 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
/linux-4.4.14/kernel/locking/
H A Dosq_lock.c38 struct optimistic_spin_node *prev) osq_wait_next()
45 * If there is a prev node in queue, then the 'old' value will be osq_wait_next()
46 * the prev node's CPU #, else it's set to OSQ_UNLOCKED_VAL since if osq_wait_next()
49 old = prev ? prev->cpu : OSQ_UNLOCKED_VAL; osq_wait_next()
55 * We were the last queued, we moved @lock back. @prev osq_wait_next()
65 * @node->next might complete Step-A and think its @prev is osq_wait_next()
87 struct optimistic_spin_node *prev, *next; osq_lock() local
105 prev = decode_cpu(old); osq_lock()
106 node->prev = prev; osq_lock()
107 WRITE_ONCE(prev->next, node); osq_lock()
110 * Normally @prev is untouchable after the above store; because at that osq_lock()
131 * Step - A -- stabilize @prev osq_lock()
133 * Undo our @prev->next assignment; this will make @prev's osq_lock()
139 if (prev->next == node && osq_lock()
140 cmpxchg(&prev->next, node, NULL) == node) osq_lock()
155 * case its step-C will write us a new @node->prev pointer. osq_lock()
157 prev = READ_ONCE(node->prev); osq_lock()
164 * back to @prev. osq_lock()
167 next = osq_wait_next(lock, node, prev); osq_lock()
174 * @prev is stable because its still waiting for a new @prev->next osq_lock()
179 WRITE_ONCE(next->prev, prev); osq_lock()
180 WRITE_ONCE(prev->next, next); osq_lock()
36 osq_wait_next(struct optimistic_spin_queue *lock, struct optimistic_spin_node *node, struct optimistic_spin_node *prev) osq_wait_next() argument
H A Dmcs_spinlock.h64 struct mcs_spinlock *prev; mcs_spin_lock() local
76 prev = xchg(lock, node); mcs_spin_lock()
77 if (likely(prev == NULL)) { mcs_spin_lock()
88 WRITE_ONCE(prev->next, node); mcs_spin_lock()
H A Dmutex.h17 __list_del((waiter)->list.prev, (waiter)->list.next)
/linux-4.4.14/arch/metag/include/asm/
H A Dmmu_context.h72 static inline void switch_mmu(struct mm_struct *prev, struct mm_struct *next) switch_mmu() argument
79 /* prev->context == prev->pgd in the case where we are initially switch_mmu()
81 if (prev->context.pgd_base != (unsigned long) prev->pgd) { switch_mmu()
83 ((pgd_t *) prev->context.pgd_base)[i] = prev->pgd[i]; switch_mmu()
85 prev->pgd = (pgd_t *)mmu_get_base(); switch_mmu()
87 next->pgd = prev->pgd; switch_mmu()
88 prev->pgd = (pgd_t *) prev->context.pgd_base; switch_mmu()
98 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
101 if (prev != next) switch_mm()
102 switch_mmu(prev, next); switch_mm()
/linux-4.4.14/arch/mips/include/asm/
H A Dswitch_to.h25 * @prev: The task previously executed.
29 * This function is used whilst scheduling to save the context of prev & load
30 * the context of next. Returns prev.
32 extern asmlinkage struct task_struct *resume(struct task_struct *prev,
52 #define __mips_mt_fpaff_switch_to(prev) \
54 struct thread_info *__prev_ti = task_thread_info(prev); \
58 (!(KSTK_STATUS(prev) & ST0_CU1))) { \
60 prev->cpus_allowed = prev->thread.user_cpus_allowed; \
66 #define __mips_mt_fpaff_switch_to(prev) do { (void) (prev); } while (0)
84 #define switch_to(prev, next, last) \
86 __mips_mt_fpaff_switch_to(prev); \
87 lose_fpu_inatomic(1, prev); \
89 __save_dsp(prev); \
94 if ((KSTK_STATUS(prev) & ST0_CU2)) { \
96 KSTK_STATUS(prev) &= ~ST0_CU2; \
97 cop2_save(prev); \
109 (last) = resume(prev, next, task_thread_info(next)); \
H A Dmmu_context.h139 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
157 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
179 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
194 cpumask_clear_cpu(cpu, mm_cpumask(prev)); activate_mm()
/linux-4.4.14/arch/avr32/include/asm/
H A Dswitch_to.h18 #define ocd_switch(prev, next) \
20 ocd_write(PID, prev->pid); \
24 #define ocd_switch(prev, next)
28 * switch_to(prev, next, last) should switch from task `prev' to task
29 * `next'. `prev' will never be the same as `next'.
41 #define switch_to(prev, next, last) \
43 ocd_switch(prev, next); \
44 last = __switch_to(prev, &prev->thread.cpu_context + 1, \
H A Dmmu_context.h108 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument
112 if (likely(prev != next)) { switch_mm()
122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
/linux-4.4.14/drivers/fmc/
H A Dfmc-dump.c25 const uint8_t *prev) dump_line()
29 if (!prev || memcmp(line, prev, LINELEN)) { dump_line()
39 if (line == prev + LINELEN) dump_line()
41 return prev; dump_line()
46 const uint8_t *line, *prev; fmc_dump_eeprom() local
58 prev = NULL; fmc_dump_eeprom()
60 prev = dump_line(i, line, prev); fmc_dump_eeprom()
65 const uint8_t *line, *prev; fmc_dump_sdb() local
96 prev = NULL; fmc_dump_sdb()
98 prev = dump_line(i, line, prev); fmc_dump_sdb()
24 dump_line(int addr, const uint8_t *line, const uint8_t *prev) dump_line() argument
/linux-4.4.14/arch/powerpc/include/asm/
H A Dcmpxchg.h18 unsigned long prev; __xchg_u32() local
27 : "=&r" (prev), "+m" (*(volatile unsigned int *)p) __xchg_u32()
31 return prev; __xchg_u32()
43 unsigned long prev; __xchg_u32_local() local
50 : "=&r" (prev), "+m" (*(volatile unsigned int *)p) __xchg_u32_local()
54 return prev; __xchg_u32_local()
61 unsigned long prev; __xchg_u64() local
70 : "=&r" (prev), "+m" (*(volatile unsigned long *)p) __xchg_u64()
74 return prev; __xchg_u64()
80 unsigned long prev; __xchg_u64_local() local
87 : "=&r" (prev), "+m" (*(volatile unsigned long *)p) __xchg_u64_local()
91 return prev; __xchg_u64_local()
151 unsigned int prev; __cmpxchg_u32() local
164 : "=&r" (prev), "+m" (*p) __cmpxchg_u32()
168 return prev; __cmpxchg_u32()
175 unsigned int prev; __cmpxchg_u32_local() local
186 : "=&r" (prev), "+m" (*p) __cmpxchg_u32_local()
190 return prev; __cmpxchg_u32_local()
197 unsigned long prev; __cmpxchg_u64() local
209 : "=&r" (prev), "+m" (*p) __cmpxchg_u64()
213 return prev; __cmpxchg_u64()
220 unsigned long prev; __cmpxchg_u64_local() local
230 : "=&r" (prev), "+m" (*p) __cmpxchg_u64_local()
234 return prev; __cmpxchg_u64_local()
H A Dswitch_to.h13 #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next)))
16 extern struct task_struct *_switch(struct thread_struct *prev,
19 static inline void save_early_sprs(struct thread_struct *prev) save_early_sprs() argument
22 prev->tar = mfspr(SPRN_TAR); save_early_sprs()
24 prev->dscr = mfspr(SPRN_DSCR); save_early_sprs()
27 static inline void save_early_sprs(struct thread_struct *prev) {} save_early_sprs() argument
H A Dmmu_context.h37 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next);
59 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
75 if (prev == next) switch_mm()
79 /* Switch coprocessor context only if prev or next uses a coprocessor */ switch_mm()
80 if (prev->context.acop || next->context.acop) switch_mm()
99 switch_mmu_context(prev, next); switch_mm()
110 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
115 switch_mm(prev, next, current); activate_mm()
H A Dfutex.h90 u32 prev; futex_atomic_cmpxchg_inatomic() local
112 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic()
116 *uval = prev; futex_atomic_cmpxchg_inatomic()
/linux-4.4.14/arch/arm/include/asm/
H A Dswitch_to.h19 * switch_to(prev, next) should switch from task `prev' to `next'
20 * `prev' will never be the same as `next'. schedule() itself
25 #define switch_to(prev,next,last) \
28 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
H A Dmmu_context.h93 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
116 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
132 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { switch_mm()
135 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
/linux-4.4.14/arch/um/os-Linux/
H A Dirq.c82 struct irq_fd **prev; os_free_irq_by_cb() local
85 prev = &active_fds; os_free_irq_by_cb()
86 while (*prev != NULL) { os_free_irq_by_cb()
87 if ((*test)(*prev, arg)) { os_free_irq_by_cb()
88 struct irq_fd *old_fd = *prev; os_free_irq_by_cb()
90 (pollfds[i].fd != (*prev)->fd)) { os_free_irq_by_cb()
94 (*prev)->fd, pollfds[i].fd); os_free_irq_by_cb()
107 *last_irq_ptr2 = prev; os_free_irq_by_cb()
109 *prev = (*prev)->next; os_free_irq_by_cb()
115 prev = &(*prev)->next; os_free_irq_by_cb()
/linux-4.4.14/arch/microblaze/include/asm/
H A Dswitch_to.h15 extern struct task_struct *_switch_to(struct thread_info *prev,
18 #define switch_to(prev, next, last) \
20 (last) = _switch_to(task_thread_info(prev), \
H A Dfutex.h101 u32 prev; futex_atomic_cmpxchg_inatomic() local
120 : "+r" (ret), "=&r" (prev), "=&r"(cmp) \ futex_atomic_cmpxchg_inatomic()
123 *uval = prev; futex_atomic_cmpxchg_inatomic()
/linux-4.4.14/arch/ia64/include/asm/
H A Dswitch_to.h44 #define __switch_to(prev,next,last) do { \
45 if (IA64_HAS_EXTRA_STATE(prev)) \
46 ia64_save_extra(prev); \
60 # define switch_to(prev,next,last) do { \
61 if (ia64_psr(task_pt_regs(prev))->mfh && ia64_is_local_fpu_owner(prev)) { \
62 ia64_psr(task_pt_regs(prev))->mfh = 0; \
63 (prev)->thread.flags |= IA64_THREAD_FPH_VALID; \
64 __ia64_save_fpu((prev)->thread.fph); \
66 __switch_to(prev, next, last); \
76 # define switch_to(prev,next,last) __switch_to(prev, next, last)
H A Dfutex.h110 unsigned long prev; futex_atomic_cmpxchg_inatomic() local
117 : "+r" (r8), "=&r" (prev) futex_atomic_cmpxchg_inatomic()
121 *uval = prev; futex_atomic_cmpxchg_inatomic()
/linux-4.4.14/arch/c6x/include/asm/
H A Dswitch_to.h20 asmlinkage void *__switch_to(struct thread_struct *prev,
24 #define switch_to(prev, next, last) \
27 (last) = __switch_to(&(prev)->thread, \
28 &(next)->thread, (prev)); \
/linux-4.4.14/include/linux/
H A Duser-return-notifier.h18 static inline void propagate_user_return_notify(struct task_struct *prev, propagate_user_return_notify() argument
21 if (test_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY)) { propagate_user_return_notify()
22 clear_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY); propagate_user_return_notify()
38 static inline void propagate_user_return_notify(struct task_struct *prev, propagate_user_return_notify() argument
H A Dvtime.h42 extern void vtime_task_switch(struct task_struct *prev); vtime_accounting_enabled()
44 extern void vtime_common_task_switch(struct task_struct *prev); vtime_task_switch()
45 static inline void vtime_task_switch(struct task_struct *prev) vtime_task_switch() argument
48 vtime_common_task_switch(prev); vtime_task_switch()
69 static inline void vtime_task_switch(struct task_struct *prev) { } vtime_account_system() argument
H A Dlist.h15 * sometimes we already know the next/prev entries and we can
28 list->prev = list; INIT_LIST_HEAD()
35 * the prev/next entries already!
39 struct list_head *prev, __list_add()
42 next->prev = new; __list_add()
44 new->prev = prev; __list_add()
45 prev->next = new; __list_add()
49 struct list_head *prev,
77 __list_add(new, head->prev, head); list_add_tail()
81 * Delete a list entry by making the prev/next entries
85 * the prev/next entries already!
87 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument
89 next->prev = prev; __list_del()
90 WRITE_ONCE(prev->next, next); __list_del()
102 __list_del(entry->prev, entry->next); __list_del_entry()
107 __list_del(entry->prev, entry->next); list_del()
109 entry->prev = LIST_POISON2; list_del()
127 new->next->prev = new; list_replace()
128 new->prev = old->prev; list_replace()
129 new->prev->next = new; list_replace()
198 * in the process of modifying either member (next or prev)
208 return (next == head) && (next == head->prev); list_empty_careful()
231 return !list_empty(head) && (head->next == head->prev); list_is_singular()
239 list->next->prev = list; __list_cut_position()
240 list->prev = entry; __list_cut_position()
243 new_first->prev = head; __list_cut_position()
275 struct list_head *prev, __list_splice()
279 struct list_head *last = list->prev; __list_splice()
281 first->prev = prev; __list_splice()
282 prev->next = first; __list_splice()
285 next->prev = last; __list_splice()
309 __list_splice(list, head->prev, head); list_splice_tail()
340 __list_splice(list, head->prev, head); list_splice_tail_init()
374 list_entry((ptr)->prev, type, member)
396 * list_prev_entry - get the prev element in list
401 list_entry((pos)->member.prev, typeof(*(pos)), member)
417 for (pos = (head)->prev; pos != (head); pos = pos->prev)
436 for (pos = (head)->prev, n = pos->prev; \
438 pos = n, n = pos->prev)
660 struct hlist_node *prev) hlist_add_behind()
662 n->next = prev->next; hlist_add_behind()
663 prev->next = n; hlist_add_behind()
664 n->pprev = &prev->next; hlist_add_behind()
38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
659 hlist_add_behind(struct hlist_node *n, struct hlist_node *prev) hlist_add_behind() argument
H A Drculist.h33 WRITE_ONCE(list->prev, list); INIT_LIST_HEAD_RCU()
46 * the prev/next entries already!
50 struct list_head *prev, struct list_head *next) __list_add_rcu()
53 new->prev = prev; __list_add_rcu()
54 rcu_assign_pointer(list_next_rcu(prev), new); __list_add_rcu()
55 next->prev = new; __list_add_rcu()
59 struct list_head *prev, struct list_head *next);
102 __list_add_rcu(new, head->prev, head); list_add_tail_rcu()
132 entry->prev = LIST_POISON2; list_del_rcu()
175 new->prev = old->prev; list_replace_rcu()
176 rcu_assign_pointer(list_next_rcu(new->prev), new); list_replace_rcu()
177 new->next->prev = new; list_replace_rcu()
178 old->prev = LIST_POISON2; list_replace_rcu()
203 struct list_head *last = list->prev; list_splice_init_rcu()
230 * permitted to traverse the prev pointers without excluding list_splice_init_rcu()
236 first->prev = head; list_splice_init_rcu()
237 at->prev = last; list_splice_init_rcu()
434 * @prev: the existing element to add the new element after.
450 struct hlist_node *prev) hlist_add_behind_rcu()
452 n->next = prev->next; hlist_add_behind_rcu()
453 n->pprev = &prev->next; hlist_add_behind_rcu()
454 rcu_assign_pointer(hlist_next_rcu(prev), n); hlist_add_behind_rcu()
49 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
449 hlist_add_behind_rcu(struct hlist_node *n, struct hlist_node *prev) hlist_add_behind_rcu() argument
H A Dosq_lock.h9 struct optimistic_spin_node *next, *prev; member in struct:optimistic_spin_node
H A Dof_pdt.h20 * If prev is NULL, the first property will be returned.
22 int (*nextprop)(phandle node, char *prev, char *buf);
H A Dagpgart.h93 struct agp_client *prev; member in struct:agp_client
101 struct agp_controller *prev; member in struct:agp_controller
116 struct agp_file_private *prev; member in struct:agp_file_private
/linux-4.4.14/arch/sh/include/asm/
H A Dswitch_to_64.h19 struct task_struct *sh64_switch_to(struct task_struct *prev,
24 #define switch_to(prev,next,last) \
30 last = sh64_switch_to(prev, &prev->thread, next, \
H A Dfutex-irq.h96 u32 prev = 0; atomic_futex_op_cmpxchg_inatomic() local
100 ret = get_user(prev, uaddr); atomic_futex_op_cmpxchg_inatomic()
101 if (!ret && oldval == prev) atomic_futex_op_cmpxchg_inatomic()
106 *uval = prev; atomic_futex_op_cmpxchg_inatomic()
H A Dswitch_to_32.h63 struct task_struct *__switch_to(struct task_struct *prev,
69 #define switch_to(prev, next, last) \
79 if (is_dsp_enabled(prev)) \
80 __save_dsp(prev); \
84 __ts1 = (u32 *)&prev->thread.sp; \
85 __ts2 = (u32 *)&prev->thread.pc; \
86 __ts4 = (u32 *)prev; \
H A Dmmu_context.h118 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument
124 if (likely(prev != next)) { switch_mm()
133 #define activate_mm(prev, next) switch_mm((prev),(next),NULL)
/linux-4.4.14/arch/m32r/include/asm/
H A Dswitch_to.h13 * switch_to(prev, next) should switch from task `prev' to `next'
14 * `prev' will never be the same as `next'.
16 * `next' and `prev' should be struct task_struct, but it isn't always defined
28 #define switch_to(prev, next, last) do { \
37 " push %1 ; store `prev' on new stack \n" \
44 : "0" (prev), \
45 "r" (&(prev->thread.sp)), "r" (&(next->thread.sp)), \
46 "r" (&(prev->thread.lr)), "r" (&(next->thread.lr)) \
H A Dmmu_context.h121 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument
128 if (prev != next) { switch_mm()
145 #define activate_mm(prev, next) \
146 switch_mm((prev), (next), NULL)
155 #define switch_mm(prev,next,tsk) do { } while (0)
157 #define activate_mm(prev,next) do { } while (0)
/linux-4.4.14/tools/usb/usbip/libsrc/
H A Dlist.h13 * sometimes we already know the next/prev entries and we can
19 struct list_head *next, *prev; member in struct:list_head
30 list->prev = list; INIT_LIST_HEAD()
37 * the prev/next entries already!
40 struct list_head *prev, __list_add()
43 next->prev = new; __list_add()
45 new->prev = prev; __list_add()
46 prev->next = new; __list_add()
63 * Delete a list entry by making the prev/next entries
67 * the prev/next entries already!
69 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument
71 next->prev = prev; __list_del()
72 prev->next = next; __list_del()
87 __list_del(entry->prev, entry->next); __list_del_entry()
92 __list_del(entry->prev, entry->next); list_del()
94 entry->prev = LIST_POISON2; list_del()
39 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
/linux-4.4.14/arch/arc/include/asm/
H A Dswitch_to.h31 #define switch_to(prev, next, last) \
33 ARC_FPU_PREV(prev, next); \
34 last = __switch_to(prev, next);\
H A Dcmpxchg.h22 unsigned long prev; __cmpxchg() local
36 : "=&r"(prev) /* Early clobber, to prevent reg reuse */ __cmpxchg()
44 return prev; __cmpxchg()
53 int prev; __cmpxchg() local
60 prev = *p; __cmpxchg()
61 if (prev == expected) __cmpxchg()
64 return prev; __cmpxchg()
/linux-4.4.14/scripts/kconfig/
H A Dlist.h24 struct list_head *next, *prev; member in struct:list_head
79 * the prev/next entries already!
82 struct list_head *prev, __list_add()
85 next->prev = _new; __list_add()
87 _new->prev = prev; __list_add()
88 prev->next = _new; __list_add()
101 __list_add(_new, head->prev, head); list_add_tail()
105 * Delete a list entry by making the prev/next entries
109 * the prev/next entries already!
111 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument
113 next->prev = prev; __list_del()
114 prev->next = next; __list_del()
127 __list_del(entry->prev, entry->next); list_del()
129 entry->prev = (struct list_head*)LIST_POISON2; list_del()
81 __list_add(struct list_head *_new, struct list_head *prev, struct list_head *next) __list_add() argument
/linux-4.4.14/arch/x86/include/asm/
H A Dcmpxchg_32.h27 u64 prev = *ptr; set_64bit() local
32 : "=m" (*ptr), "+A" (prev) set_64bit()
48 u64 prev; __cmpxchg64() local
50 : "=A" (prev), __cmpxchg64()
56 return prev; __cmpxchg64()
61 u64 prev; __cmpxchg64_local() local
63 : "=A" (prev), __cmpxchg64_local()
69 return prev; __cmpxchg64_local()
H A Dswitch_to.h5 __visible struct task_struct *__switch_to(struct task_struct *prev,
31 #define switch_to(prev, next, last) \
55 : [prev_sp] "=m" (prev->thread.sp), \
56 [prev_ip] "=m" (prev->thread.ip), \
70 [prev] "a" (prev), \
108 #define switch_to(prev, next, last) \
110 "movq %%rsp,%P[threadrsp](%[prev])\n\t" /* save RSP */ \
122 : [next] "S" (next), [prev] "D" (prev), \
H A Dmmu_context.h15 static inline void paravirt_activate_mm(struct mm_struct *prev, paravirt_activate_mm() argument
107 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
112 if (likely(prev != next)) { switch_mm()
152 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
161 * It's possible that prev->context.ldt doesn't match switch_mm()
162 * the LDT register. This can happen if leave_mm(prev) switch_mm()
164 * prev->context.ldt but suppressed an IPI to this CPU. switch_mm()
165 * In this case, prev->context.ldt != NULL, because we switch_mm()
168 * prev->context.ldt, because mms never share an LDT. switch_mm()
170 if (unlikely(prev->context.ldt != next->context.ldt)) switch_mm()
205 #define activate_mm(prev, next) \
207 paravirt_activate_mm((prev), (next)); \
208 switch_mm((prev), (next), NULL); \
/linux-4.4.14/arch/xtensa/include/asm/
H A Dnommu_context.h18 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
22 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
H A Dswitch_to.h17 #define switch_to(prev,next,last) \
19 (last) = _switch_to(prev, next); \
H A Dfutex.h112 u32 prev; futex_atomic_cmpxchg_inatomic() local
138 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic()
142 *uval = prev; futex_atomic_cmpxchg_inatomic()
H A Dmmu_context.h123 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
133 if (migrated || prev != next) switch_mm()
137 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
/linux-4.4.14/scripts/gdb/linux/
H A Dlists.py34 p = c['prev'].dereference()
38 gdb.write('prev.next != current: '
40 'prev@{p_addr}={p}\n'.format(
48 gdb.write('prev is not accessible: '
55 if n['prev'] != c.address:
56 gdb.write('next.prev != current: '
/linux-4.4.14/arch/parisc/lib/
H A Dbitops.c62 unsigned long prev; __cmpxchg_u64() local
65 if ((prev = *ptr) == old) __cmpxchg_u64()
68 return prev; __cmpxchg_u64()
75 unsigned int prev; __cmpxchg_u32() local
78 if ((prev = *ptr) == old) __cmpxchg_u32()
81 return (unsigned long)prev; __cmpxchg_u32()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/core/
H A Dmm.c53 struct nvkm_mm_node *prev = node(this, prev); nvkm_mm_free() local
56 if (prev && prev->type == NVKM_MM_TYPE_NONE) { nvkm_mm_free()
57 prev->length += this->length; nvkm_mm_free()
59 kfree(this); this = prev; nvkm_mm_free()
72 list_for_each_entry(prev, &mm->free, fl_entry) { nvkm_mm_free()
73 if (this->offset < prev->offset) nvkm_mm_free()
77 list_add_tail(&this->fl_entry, &prev->fl_entry); nvkm_mm_free()
114 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_head() local
129 prev = node(this, prev); nvkm_mm_head()
130 if (prev && prev->type != type) nvkm_mm_head()
188 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_tail() local
202 prev = node(this, prev); nvkm_mm_tail()
203 if (prev && prev->type != type) nvkm_mm_tail()
240 struct nvkm_mm_node *node, *prev; nvkm_mm_init() local
244 prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); nvkm_mm_init()
245 next = prev->offset + prev->length; nvkm_mm_init()
/linux-4.4.14/drivers/net/fddi/skfp/
H A Dsmttimer.c41 struct smt_timer **prev ; smt_timer_stop() local
51 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { smt_timer_stop()
53 *prev = tm->tm_next ; smt_timer_stop()
65 struct smt_timer **prev ; smt_timer_start() local
92 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { smt_timer_start()
99 *prev = timer ; smt_timer_start()
/linux-4.4.14/net/bridge/
H A Dbr_forward.c24 static int deliver_clone(const struct net_bridge_port *prev,
156 static int deliver_clone(const struct net_bridge_port *prev, deliver_clone() argument
169 __packet_hook(prev, skb); deliver_clone()
174 struct net_bridge_port *prev, struct net_bridge_port *p, maybe_deliver()
182 return prev; maybe_deliver()
184 if (!prev) maybe_deliver()
187 err = deliver_clone(prev, skb, __packet_hook); maybe_deliver()
203 struct net_bridge_port *prev; br_flood() local
205 prev = NULL; br_flood()
219 prev = maybe_deliver(prev, p, skb, __packet_hook); br_flood()
220 if (IS_ERR(prev)) br_flood()
224 if (!prev) br_flood()
228 deliver_clone(prev, skb, __packet_hook); br_flood()
230 __packet_hook(prev, skb); br_flood()
262 struct net_bridge_port *prev = NULL; br_multicast_flood() local
278 prev = maybe_deliver(prev, port, skb, __packet_hook); br_multicast_flood()
279 if (IS_ERR(prev)) br_multicast_flood()
288 if (!prev) br_multicast_flood()
292 deliver_clone(prev, skb, __packet_hook); br_multicast_flood()
294 __packet_hook(prev, skb); br_multicast_flood()
173 maybe_deliver( struct net_bridge_port *prev, struct net_bridge_port *p, struct sk_buff *skb, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) maybe_deliver() argument
/linux-4.4.14/drivers/gpu/drm/radeon/
H A Dmkregtable.c36 * sometimes we already know the next/prev entries and we can
42 struct list_head *next, *prev; member in struct:list_head
53 list->prev = list; INIT_LIST_HEAD()
60 * the prev/next entries already!
64 struct list_head *prev, struct list_head *next) __list_add()
66 next->prev = new; __list_add()
68 new->prev = prev; __list_add()
69 prev->next = new; __list_add()
73 struct list_head *prev, struct list_head *next);
99 __list_add(new, head->prev, head); list_add_tail()
103 * Delete a list entry by making the prev/next entries
107 * the prev/next entries already!
109 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument
111 next->prev = prev; __list_del()
112 prev->next = next; __list_del()
124 __list_del(entry->prev, entry->next); list_del()
126 entry->prev = (void *)0xBEEFDEAD; list_del()
142 new->next->prev = new; list_replace()
143 new->prev = old->prev; list_replace()
144 new->prev->next = new; list_replace()
160 __list_del(entry->prev, entry->next); list_del_init()
171 __list_del(list->prev, list->next); list_move()
183 __list_del(list->prev, list->next); list_move_tail()
213 * in the process of modifying either member (next or prev)
223 return (next == head) && (next == head->prev); list_empty_careful()
232 return !list_empty(head) && (head->next == head->prev); list_is_singular()
241 list->next->prev = list; __list_cut_position()
242 list->prev = entry; __list_cut_position()
245 new_first->prev = head; __list_cut_position()
277 struct list_head *prev, struct list_head *next) __list_splice()
280 struct list_head *last = list->prev; __list_splice()
282 first->prev = prev; __list_splice()
283 prev->next = first; __list_splice()
286 next->prev = last; __list_splice()
310 __list_splice(list, head->prev, head); list_splice_tail()
341 __list_splice(list, head->prev, head); list_splice_tail_init()
381 for (pos = (head)->prev; prefetch(pos->prev), pos != (head); \
382 pos = pos->prev)
401 for (pos = (head)->prev, n = pos->prev; \
402 prefetch(pos->prev), pos != (head); \
403 pos = n, n = pos->prev)
423 for (pos = list_entry((head)->prev, typeof(*pos), member); \
424 prefetch(pos->member.prev), &pos->member != (head); \
425 pos = list_entry(pos->member.prev, typeof(*pos), member))
462 for (pos = list_entry(pos->member.prev, typeof(*pos), member); \
463 prefetch(pos->member.prev), &pos->member != (head); \
464 pos = list_entry(pos->member.prev, typeof(*pos), member))
533 for (pos = list_entry((head)->prev, typeof(*pos), member), \
534 n = list_entry(pos->member.prev, typeof(*pos), member); \
536 pos = n, n = list_entry(n->member.prev, typeof(*n), member))
63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
276 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
H A Dradeon_mem.c55 newblock->prev = p; split_block()
56 p->next->prev = newblock; split_block()
72 newblock->prev = p; split_block()
73 p->next->prev = newblock; split_block()
121 p->next->prev = p; free_block()
125 if (p->prev->file_priv == NULL) { free_block()
126 struct mem_block *q = p->prev; free_block()
129 q->next->prev = q; free_block()
152 blocks->next = blocks->prev = *heap; init_heap()
155 (*heap)->next = (*heap)->prev = blocks; init_heap()
181 p->next->prev = p; list_for_each()
/linux-4.4.14/arch/cris/arch-v32/mm/
H A Dintmem.c107 struct intmem_allocation *prev = crisv32_intmem_free() local
108 list_entry(allocation->entry.prev, crisv32_intmem_free()
115 /* Join with prev and/or next if also free */ crisv32_intmem_free()
116 if ((prev != &intmem_allocations) && crisv32_intmem_free()
117 (prev->status == STATUS_FREE)) { crisv32_intmem_free()
118 prev->size += allocation->size; crisv32_intmem_free()
121 allocation = prev; crisv32_intmem_free()
/linux-4.4.14/mm/
H A Dinterval_tree.c28 /* Insert node immediately after prev in the interval tree */ vma_interval_tree_insert_after()
30 struct vm_area_struct *prev, vma_interval_tree_insert_after()
37 VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); vma_interval_tree_insert_after()
39 if (!prev->shared.rb.rb_right) { vma_interval_tree_insert_after()
40 parent = prev; vma_interval_tree_insert_after()
41 link = &prev->shared.rb.rb_right; vma_interval_tree_insert_after()
43 parent = rb_entry(prev->shared.rb.rb_right, vma_interval_tree_insert_after()
29 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root *root) vma_interval_tree_insert_after() argument
H A Dmadvise.c47 struct vm_area_struct **prev, madvise_behavior()
100 *prev = vma; madvise_behavior()
105 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior()
108 if (*prev) { madvise_behavior()
109 vma = *prev; madvise_behavior()
113 *prev = vma; madvise_behavior()
221 struct vm_area_struct **prev, madvise_willneed()
228 *prev = vma; madvise_willneed()
234 *prev = vma; madvise_willneed()
249 *prev = vma; madvise_willneed()
279 struct vm_area_struct **prev, madvise_dontneed()
282 *prev = vma; madvise_dontneed()
295 struct vm_area_struct **prev, madvise_remove()
302 *prev = NULL; /* tell sys_madvise we drop mmap_sem */ madvise_remove()
374 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, madvise_vma() argument
379 return madvise_remove(vma, prev, start, end); madvise_vma()
381 return madvise_willneed(vma, prev, start, end); madvise_vma()
383 return madvise_dontneed(vma, prev, start, end); madvise_vma()
385 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma()
463 struct vm_area_struct *vma, *prev; SYSCALL_DEFINE3() local
504 vma = find_vma_prev(current->mm, start, &prev); SYSCALL_DEFINE3()
506 prev = vma; SYSCALL_DEFINE3()
529 error = madvise_vma(vma, &prev, start, tmp, behavior); SYSCALL_DEFINE3()
533 if (prev && start < prev->vm_end) SYSCALL_DEFINE3()
534 start = prev->vm_end; SYSCALL_DEFINE3()
538 if (prev) SYSCALL_DEFINE3()
539 vma = prev->vm_next; SYSCALL_DEFINE3()
46 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument
220 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument
278 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument
294 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument
H A Dslob.c219 slob_t *prev, *cur, *aligned = NULL; slob_page_alloc() local
222 for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) { slob_page_alloc()
236 prev = cur; slob_page_alloc()
243 if (prev) slob_page_alloc()
244 set_slob(prev, slob_units(prev), next); slob_page_alloc()
248 if (prev) slob_page_alloc()
249 set_slob(prev, slob_units(prev), cur + units); slob_page_alloc()
271 struct list_head *prev; slob_alloc() local
299 prev = sp->lru.prev; list_for_each_entry()
307 if (prev != slob_list->prev && list_for_each_entry()
308 slob_list->next != prev->next) list_for_each_entry()
309 list_move_tail(slob_list, prev->next); list_for_each_entry()
343 slob_t *prev, *next, *b = (slob_t *)block; slob_free() local
399 prev = sp->freelist; slob_free()
400 next = slob_next(prev); slob_free()
402 prev = next; slob_free()
403 next = slob_next(prev); slob_free()
406 if (!slob_last(prev) && b + units == next) { slob_free()
412 if (prev + slob_units(prev) == b) { slob_free()
413 units = slob_units(b) + slob_units(prev); slob_free()
414 set_slob(prev, units, slob_next(b)); slob_free()
416 set_slob(prev, slob_units(prev), b); slob_free()
H A Dmlock.c498 static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev, mlock_fixup() argument
513 *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma, mlock_fixup()
516 if (*prev) { mlock_fixup()
517 vma = *prev; mlock_fixup()
554 *prev = vma; mlock_fixup()
562 struct vm_area_struct * vma, * prev; apply_vma_lock_flags() local
576 prev = vma->vm_prev; apply_vma_lock_flags()
578 prev = vma; apply_vma_lock_flags()
589 error = mlock_fixup(vma, &prev, nstart, tmp, newflags); apply_vma_lock_flags()
593 if (nstart < prev->vm_end) apply_vma_lock_flags()
594 nstart = prev->vm_end; apply_vma_lock_flags()
598 vma = prev->vm_next; apply_vma_lock_flags()
687 struct vm_area_struct * vma, * prev = NULL; apply_mlockall_flags() local
707 for (vma = current->mm->mmap; vma ; vma = prev->vm_next) { apply_mlockall_flags()
714 mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags); apply_mlockall_flags()
H A Dmmap.c62 struct vm_area_struct *vma, struct vm_area_struct *prev,
382 unsigned long prev = 0, pend = 0; browse_rb() local
387 if (vma->vm_start < prev) { browse_rb()
388 pr_emerg("vm_start %lx < prev %lx\n", browse_rb()
389 vma->vm_start, prev); browse_rb()
410 prev = vma->vm_start; browse_rb()
661 struct vm_area_struct *prev, struct rb_node **rb_link, __vma_link()
664 __vma_link_list(mm, vma, prev, rb_parent); __vma_link()
669 struct vm_area_struct *prev, struct rb_node **rb_link, vma_link()
679 __vma_link(mm, vma, prev, rb_link, rb_parent); vma_link()
695 struct vm_area_struct *prev; __insert_vm_struct() local
699 &prev, &rb_link, &rb_parent)) __insert_vm_struct()
701 __vma_link(mm, vma, prev, rb_link, rb_parent); __insert_vm_struct()
707 struct vm_area_struct *prev) __vma_unlink()
712 prev->vm_next = next = vma->vm_next; __vma_unlink()
714 next->vm_prev = prev; __vma_unlink()
1019 * an offset within prev, or at the start of next), and the flags of
1025 * vma, PPPPPP is the prev vma specified, and NNNNNN the next vma after:
1041 struct vm_area_struct *prev, unsigned long addr, vma_merge()
1058 if (prev) vma_merge()
1059 next = prev->vm_next; vma_merge()
1069 if (prev && prev->vm_end == addr && vma_merge()
1070 mpol_equal(vma_policy(prev), policy) && vma_merge()
1071 can_vma_merge_after(prev, vm_flags, vma_merge()
1083 is_mergeable_anon_vma(prev->anon_vma, vma_merge()
1086 err = vma_adjust(prev, prev->vm_start, vma_merge()
1087 next->vm_end, prev->vm_pgoff, NULL); vma_merge()
1089 err = vma_adjust(prev, prev->vm_start, vma_merge()
1090 end, prev->vm_pgoff, NULL); vma_merge()
1093 khugepaged_enter_vma_merge(prev, vm_flags); vma_merge()
1094 return prev; vma_merge()
1105 if (prev && addr < prev->vm_end) /* case 4 */ vma_merge()
1106 err = vma_adjust(prev, prev->vm_start, vma_merge()
1107 addr, prev->vm_pgoff, NULL); vma_merge()
1545 struct vm_area_struct *vma, *prev; mmap_region() local
1568 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, mmap_region()
1587 vma = vma_merge(mm, prev, addr, addr + len, vm_flags, mmap_region()
1637 * Bug: If addr is changed, prev, rb_link, rb_parent should mmap_region()
1650 vma_link(mm, vma, prev, rb_link, rb_parent); mmap_region()
1692 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); mmap_region()
1776 struct rb_node *prev = &vma->vm_rb; unmapped_area() local
1777 if (!rb_parent(prev)) unmapped_area()
1779 vma = rb_entry(rb_parent(prev), unmapped_area()
1781 if (prev == vma->vm_rb.rb_left) { unmapped_area()
1879 struct rb_node *prev = &vma->vm_rb; unmapped_area_topdown() local
1880 if (!rb_parent(prev)) unmapped_area_topdown()
1882 vma = rb_entry(rb_parent(prev), unmapped_area_topdown()
1884 if (prev == vma->vm_rb.rb_right) { unmapped_area_topdown()
2320 struct vm_area_struct *vma, *prev; find_extend_vma() local
2323 vma = find_vma_prev(mm, addr, &prev); find_extend_vma()
2326 if (!prev || expand_stack(prev, addr)) find_extend_vma()
2328 if (prev->vm_flags & VM_LOCKED) find_extend_vma()
2329 populate_vma_page_range(prev, addr, prev->vm_end, NULL); find_extend_vma()
2330 return prev; find_extend_vma()
2335 struct vm_area_struct *prev; expand_stack() local
2338 prev = vma->vm_prev; expand_stack()
2339 if (prev && prev->vm_end == address) { expand_stack()
2340 if (!(prev->vm_flags & VM_GROWSDOWN)) expand_stack()
2401 struct vm_area_struct *vma, struct vm_area_struct *prev, unmap_region()
2404 struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; unmap_region()
2411 free_pgtables(&tlb, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, unmap_region()
2422 struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped()
2427 insertion_point = (prev ? &prev->vm_next : &mm->mmap); detach_vmas_to_be_unmapped()
2437 vma->vm_prev = prev; detach_vmas_to_be_unmapped()
2440 mm->highest_vm_end = prev ? prev->vm_end : 0; detach_vmas_to_be_unmapped()
2535 struct vm_area_struct *vma, *prev, *last; do_munmap() local
2548 prev = vma->vm_prev; do_munmap()
2577 prev = vma; do_munmap()
2587 vma = prev ? prev->vm_next : mm->mmap; do_munmap()
2606 detach_vmas_to_be_unmapped(mm, vma, prev, end); do_munmap()
2607 unmap_region(mm, vma, prev, start, end); do_munmap()
2746 struct vm_area_struct *vma, *prev; do_brk() local
2775 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, do_brk()
2792 vma = vma_merge(mm, prev, addr, addr + len, flags, do_brk()
2813 vma_link(mm, vma, prev, rb_link, rb_parent); do_brk()
2892 struct vm_area_struct *prev; insert_vm_struct() local
2896 &prev, &rb_link, &rb_parent)) insert_vm_struct()
2919 vma_link(mm, vma, prev, rb_link, rb_parent); insert_vm_struct()
2934 struct vm_area_struct *new_vma, *prev; copy_vma() local
2947 if (find_vma_links(mm, addr, addr + len, &prev, &rb_link, &rb_parent)) copy_vma()
2949 new_vma = vma_merge(mm, prev, addr, addr + len, vma->vm_flags, copy_vma()
2991 vma_link(mm, new_vma, prev, rb_link, rb_parent); copy_vma()
660 __vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) __vma_link() argument
668 vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) vma_link() argument
706 __vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev) __vma_unlink() argument
1040 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy, struct vm_userfaultfd_ctx vm_userfaultfd_ctx) vma_merge() argument
2400 unmap_region(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long start, unsigned long end) unmap_region() argument
2421 detach_vmas_to_be_unmapped(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped() argument
H A Dmprotect.c351 struct vm_area_struct *vma, *prev; SYSCALL_DEFINE3() local
384 prev = vma->vm_prev; SYSCALL_DEFINE3()
403 prev = vma; SYSCALL_DEFINE3()
426 error = mprotect_fixup(vma, &prev, nstart, tmp, newflags); SYSCALL_DEFINE3()
431 if (nstart < prev->vm_end) SYSCALL_DEFINE3()
432 nstart = prev->vm_end; SYSCALL_DEFINE3()
436 vma = prev->vm_next; SYSCALL_DEFINE3()
/linux-4.4.14/sound/pci/ctxfi/
H A Dctimap.c46 pre = pos->prev;
48 pre = head->prev;
50 __list_add(&entry->list, pos->prev, pos);
52 pre = head->prev;
79 pre = (entry->list.prev == head) ? head->prev : entry->list.prev; input_mapper_delete()
/linux-4.4.14/arch/alpha/include/asm/
H A Dxchg.h132 * prev is equal to old) then we aren't acquiring anything new and
139 unsigned long prev, tmp, cmp, addr64; ____cmpxchg() local
157 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) ____cmpxchg()
160 return prev; ____cmpxchg()
166 unsigned long prev, tmp, cmp, addr64; ____cmpxchg() local
184 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) ____cmpxchg()
187 return prev; ____cmpxchg()
193 unsigned long prev, cmp; ____cmpxchg() local
207 : "=&r"(prev), "=&r"(cmp), "=m"(*m) ____cmpxchg()
210 return prev; ____cmpxchg()
216 unsigned long prev, cmp; ____cmpxchg() local
230 : "=&r"(prev), "=&r"(cmp), "=m"(*m) ____cmpxchg()
233 return prev; ____cmpxchg()
H A Dfutex.h88 u32 prev; futex_atomic_cmpxchg_inatomic() local
110 : "+r"(ret), "=&r"(prev), "=&r"(cmp) futex_atomic_cmpxchg_inatomic()
114 *uval = prev; futex_atomic_cmpxchg_inatomic()
H A Dgct.h24 gct6_handle prev; member in struct:__gct6_node
/linux-4.4.14/arch/avr32/kernel/
H A Dswitch_to.S15 /* Switch thread context from "prev" to "next", returning "last"
16 * r12 : prev
17 * r11 : &prev->thread + 1
/linux-4.4.14/tools/include/linux/
H A Dlist.h19 begin->prev->next = end->next; list_del_range()
20 end->next->prev = begin->prev; list_del_range()
/linux-4.4.14/security/selinux/ss/
H A Dhashtab.c42 struct hashtab_node *prev, *cur, *newnode; hashtab_insert() local
50 prev = NULL; hashtab_insert()
53 prev = cur; hashtab_insert()
65 if (prev) { hashtab_insert()
66 newnode->next = prev->next; hashtab_insert()
67 prev->next = newnode; hashtab_insert()
H A Debitmap.c51 struct ebitmap_node *n, *new, *prev; ebitmap_cpy() local
55 prev = NULL; ebitmap_cpy()
65 if (prev) ebitmap_cpy()
66 prev->next = new; ebitmap_cpy()
69 prev = new; ebitmap_cpy()
258 struct ebitmap_node *n, *prev, *new; ebitmap_set_bit() local
260 prev = NULL; ebitmap_set_bit()
281 if (prev) ebitmap_set_bit()
282 e->highbit = prev->startbit ebitmap_set_bit()
287 if (prev) ebitmap_set_bit()
288 prev->next = n->next; ebitmap_set_bit()
295 prev = n; ebitmap_set_bit()
313 if (prev) { ebitmap_set_bit()
314 new->next = prev->next; ebitmap_set_bit()
315 prev->next = new; ebitmap_set_bit()
H A Dmls.c37 int i, l, len, head, prev; mls_compute_context_len() local
52 prev = -2; mls_compute_context_len()
55 if (i - prev > 1) { ebitmap_for_each_positive_bit()
57 if (head != prev) { ebitmap_for_each_positive_bit()
58 nm = sym_name(&policydb, SYM_CATS, prev); ebitmap_for_each_positive_bit()
65 prev = i; ebitmap_for_each_positive_bit()
67 if (prev != head) {
68 nm = sym_name(&policydb, SYM_CATS, prev);
92 int i, l, head, prev; mls_sid_to_context() local
111 prev = -2; mls_sid_to_context()
114 if (i - prev > 1) { ebitmap_for_each_positive_bit()
116 if (prev != head) { ebitmap_for_each_positive_bit()
117 if (prev - head > 1) ebitmap_for_each_positive_bit()
121 nm = sym_name(&policydb, SYM_CATS, prev); ebitmap_for_each_positive_bit()
125 if (prev < 0) ebitmap_for_each_positive_bit()
134 prev = i; ebitmap_for_each_positive_bit()
137 if (prev != head) {
138 if (prev - head > 1)
142 nm = sym_name(&policydb, SYM_CATS, prev);
H A Dsidtab.c36 struct sidtab_node *prev, *cur, *newnode; sidtab_insert() local
44 prev = NULL; sidtab_insert()
47 prev = cur; sidtab_insert()
68 if (prev) { sidtab_insert()
69 newnode->next = prev->next; sidtab_insert()
71 prev->next = newnode; sidtab_insert()
/linux-4.4.14/fs/jffs2/
H A Dnodelist.c28 struct jffs2_full_dirent **prev = list; jffs2_add_fd_to_list() local
32 while ((*prev) && (*prev)->nhash <= new->nhash) { jffs2_add_fd_to_list()
33 if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { jffs2_add_fd_to_list()
35 if (new->version < (*prev)->version) { jffs2_add_fd_to_list()
37 (*prev)->name, (*prev)->ino); jffs2_add_fd_to_list()
42 (*prev)->name, (*prev)->ino); jffs2_add_fd_to_list()
43 new->next = (*prev)->next; jffs2_add_fd_to_list()
46 if ((*prev)->raw) jffs2_add_fd_to_list()
47 jffs2_mark_node_obsolete(c, ((*prev)->raw)); jffs2_add_fd_to_list()
48 jffs2_free_full_dirent(*prev); jffs2_add_fd_to_list()
49 *prev = new; jffs2_add_fd_to_list()
53 prev = &((*prev)->next); jffs2_add_fd_to_list()
55 new->next = *prev; jffs2_add_fd_to_list()
56 *prev = new; jffs2_add_fd_to_list()
386 struct jffs2_node_frag *prev = frag_prev(newfrag); jffs2_add_full_dnode_to_inode() local
390 if (prev->node) jffs2_add_full_dnode_to_inode()
391 mark_ref_normal(prev->node->raw); jffs2_add_full_dnode_to_inode()
438 struct jffs2_inode_cache **prev; jffs2_add_ino_cache() local
446 prev = &c->inocache_list[new->ino % c->inocache_hashsize]; jffs2_add_ino_cache()
448 while ((*prev) && (*prev)->ino < new->ino) { jffs2_add_ino_cache()
449 prev = &(*prev)->next; jffs2_add_ino_cache()
451 new->next = *prev; jffs2_add_ino_cache()
452 *prev = new; jffs2_add_ino_cache()
459 struct jffs2_inode_cache **prev; jffs2_del_ino_cache() local
467 prev = &c->inocache_list[old->ino % c->inocache_hashsize]; jffs2_del_ino_cache()
469 while ((*prev) && (*prev)->ino < old->ino) { jffs2_del_ino_cache()
470 prev = &(*prev)->next; jffs2_del_ino_cache()
472 if ((*prev) == old) { jffs2_del_ino_cache()
473 *prev = old->next; jffs2_del_ino_cache()
529 struct jffs2_node_frag *prev = NULL; jffs2_lookup_node_frag() local
541 if (!prev || frag->ofs > prev->ofs) jffs2_lookup_node_frag()
542 prev = frag; jffs2_lookup_node_frag()
554 if (prev) jffs2_lookup_node_frag()
556 prev->ofs, prev->ofs+prev->size); jffs2_lookup_node_frag()
560 return prev; jffs2_lookup_node_frag()
H A Derase.c239 struct jffs2_raw_node_ref **prev; jffs2_remove_node_refs_from_ino_list() local
241 prev = &ref->next_in_ino; jffs2_remove_node_refs_from_ino_list()
245 if (!(*prev)->next_in_ino) { jffs2_remove_node_refs_from_ino_list()
249 ic = (struct jffs2_inode_cache *)(*prev); jffs2_remove_node_refs_from_ino_list()
250 prev = &ic->nodes; jffs2_remove_node_refs_from_ino_list()
254 if (SECTOR_ADDR((*prev)->flash_offset) == jeb->offset) { jffs2_remove_node_refs_from_ino_list()
258 this = *prev; jffs2_remove_node_refs_from_ino_list()
259 *prev = this->next_in_ino; jffs2_remove_node_refs_from_ino_list()
268 prev = &((*prev)->next_in_ino); jffs2_remove_node_refs_from_ino_list()
/linux-4.4.14/drivers/video/fbdev/omap2/dss/
H A Ddss-of.c27 struct device_node *prev) omapdss_of_get_next_port()
34 if (!prev) { omapdss_of_get_next_port()
51 ports = of_get_parent(prev); omapdss_of_get_next_port()
56 port = of_get_next_child(ports, prev); omapdss_of_get_next_port()
61 prev = port; omapdss_of_get_next_port()
73 struct device_node *prev) omapdss_of_get_next_endpoint()
81 ep = of_get_next_child(parent, prev); omapdss_of_get_next_endpoint()
84 prev = ep; omapdss_of_get_next_endpoint()
26 omapdss_of_get_next_port(const struct device_node *parent, struct device_node *prev) omapdss_of_get_next_port() argument
72 omapdss_of_get_next_endpoint(const struct device_node *parent, struct device_node *prev) omapdss_of_get_next_endpoint() argument
/linux-4.4.14/arch/nios2/include/asm/
H A Dswitch_to.h17 #define switch_to(prev, next, last) \
26 : "r" (prev), "r" (next) \
H A Dmmu_context.h52 void switch_mm(struct mm_struct *prev, struct mm_struct *next,
64 void activate_mm(struct mm_struct *prev, struct mm_struct *next);
/linux-4.4.14/fs/btrfs/
H A Dextent_map.c145 struct rb_node *prev = NULL; __tree_search() local
152 prev = n; __tree_search()
164 orig_prev = prev; __tree_search()
165 while (prev && offset >= extent_map_end(prev_entry)) { __tree_search()
166 prev = rb_next(prev); __tree_search()
167 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search()
169 *prev_ret = prev; __tree_search()
170 prev = orig_prev; __tree_search()
174 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search()
175 while (prev && offset < prev_entry->start) { __tree_search()
176 prev = rb_prev(prev); __tree_search()
177 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search()
179 *next_ret = prev; __tree_search()
185 static int mergable_maps(struct extent_map *prev, struct extent_map *next) mergable_maps() argument
187 if (test_bit(EXTENT_FLAG_PINNED, &prev->flags)) mergable_maps()
194 if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags)) mergable_maps()
197 if (test_bit(EXTENT_FLAG_LOGGING, &prev->flags) || mergable_maps()
206 if (!list_empty(&prev->list) || !list_empty(&next->list)) mergable_maps()
209 if (extent_map_end(prev) == next->start && mergable_maps()
210 prev->flags == next->flags && mergable_maps()
211 prev->bdev == next->bdev && mergable_maps()
213 prev->block_start == EXTENT_MAP_HOLE) || mergable_maps()
215 prev->block_start == EXTENT_MAP_INLINE) || mergable_maps()
217 prev->block_start == EXTENT_MAP_DELALLOC) || mergable_maps()
219 next->block_start == extent_map_block_end(prev)))) { mergable_maps()
365 struct rb_node *prev = NULL; __lookup_extent_mapping() local
369 rb_node = __tree_search(&tree->map, start, &prev, &next); __lookup_extent_mapping()
371 if (prev) __lookup_extent_mapping()
372 rb_node = prev; __lookup_extent_mapping()
H A Dordered-data.c81 struct rb_node *prev = NULL; __tree_search() local
88 prev = n; __tree_search()
101 while (prev && file_offset >= entry_end(prev_entry)) { __tree_search()
102 test = rb_next(prev); __tree_search()
110 prev = test; __tree_search()
112 if (prev) __tree_search()
113 prev_entry = rb_entry(prev, struct btrfs_ordered_extent, __tree_search()
115 while (prev && file_offset < entry_end(prev_entry)) { __tree_search()
116 test = rb_prev(prev); __tree_search()
121 prev = test; __tree_search()
123 *prev_ret = prev; __tree_search()
155 struct rb_node *prev = NULL; tree_search() local
165 ret = __tree_search(root, file_offset, &prev); tree_search()
167 ret = prev; tree_search()
441 struct rb_node *prev; btrfs_get_logged_extents() local
445 n = __tree_search(&tree->tree, end, &prev); btrfs_get_logged_extents()
447 n = prev; btrfs_get_logged_extents()
951 struct rb_node *prev = NULL; btrfs_ordered_update_i_size() local
997 prev = tree_search(tree, offset); btrfs_ordered_update_i_size()
1002 if (prev) { btrfs_ordered_update_i_size()
1003 test = rb_entry(prev, struct btrfs_ordered_extent, btrfs_ordered_update_i_size()
1007 node = prev; btrfs_ordered_update_i_size()
/linux-4.4.14/include/uapi/linux/netfilter_bridge/
H A Debt_limit.h19 unsigned long prev; member in struct:ebt_limit_info
/linux-4.4.14/arch/mips/kernel/
H A Dcsrc-r4k.c47 unsigned int prev, curr, i; rdhwr_count_usable() local
55 for (i = 0, prev = rdhwr_count(); i < 100; i++) { rdhwr_count_usable()
58 if (curr != prev) rdhwr_count_usable()
61 prev = curr; rdhwr_count_usable()
/linux-4.4.14/arch/h8300/include/asm/
H A Dswitch_to.h27 * pass prev in a0, next in a1, offset of tss in d1, and whether
34 #define switch_to(prev, next, last) \
44 : "r" (&(prev->thread)), \
46 "g" (prev) \
/linux-4.4.14/drivers/usb/host/
H A Dohci-mem.c105 struct td **prev = &hc->td_hash [TD_HASH_FUNC (td->td_dma)]; td_free() local
107 while (*prev && *prev != td) td_free()
108 prev = &(*prev)->td_hash; td_free()
109 if (*prev) td_free()
110 *prev = td->td_hash; td_free()
/linux-4.4.14/arch/m68k/include/asm/
H A Datomic.h132 int prev; atomic_cmpxchg() local
135 prev = atomic_read(v); atomic_cmpxchg()
136 if (prev == old) atomic_cmpxchg()
139 return prev; atomic_cmpxchg()
145 int prev; atomic_xchg() local
148 prev = atomic_read(v); atomic_xchg()
151 return prev; atomic_xchg()
H A Dswitch_to.h27 * pass prev in a0, next in a1
30 #define switch_to(prev,next,last) do { \
31 register void *_prev __asm__ ("a0") = (prev); \
/linux-4.4.14/drivers/gpu/drm/nouveau/include/nvif/
H A Dlist.h111 struct list_head *next, *prev; member in struct:list_head
130 list->next = list->prev = list; INIT_LIST_HEAD()
135 struct list_head *prev, struct list_head *next) __list_add()
137 next->prev = entry; __list_add()
139 entry->prev = prev; __list_add()
140 prev->next = entry; __list_add()
182 __list_add(entry, head->prev, head); list_add_tail()
186 __list_del(struct list_head *prev, struct list_head *next) __list_del() argument
188 next->prev = prev; __list_del()
189 prev->next = next; __list_del()
209 __list_del(entry->prev, entry->next); list_del()
215 __list_del(entry->prev, entry->next); list_del_init()
222 __list_del(list->prev, list->next); list_move_tail()
292 list_entry((ptr)->prev, type, member)
334 for (pos = __container_of((head)->prev, pos, member); \
336 pos = __container_of(pos->member.prev, pos, member))
344 for (pos = __container_of(pos->member.prev, pos, member); \
346 pos = __container_of(pos->member.prev, pos, member))
134 __list_add(struct list_head *entry, struct list_head *prev, struct list_head *next) __list_add() argument
/linux-4.4.14/arch/tile/include/asm/
H A Dswitch_to.h40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next)))
41 extern struct task_struct *_switch_to(struct task_struct *prev,
45 extern struct task_struct *__switch_to(struct task_struct *prev,
73 struct task_struct *sim_notify_fork(struct task_struct *prev);
H A Dmmu_context.h91 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
94 if (likely(prev != next)) { switch_mm()
107 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
114 check_mm_caching(prev, next); switch_mm()
/linux-4.4.14/arch/hexagon/include/asm/
H A Dmmu_context.h66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
75 if (next->context.generation < prev->context.generation) { switch_mm()
79 next->context.generation = prev->context.generation; switch_mm()
88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
93 switch_mm(prev, next, current_thread_info()->task); activate_mm()
H A Dfutex.h106 int prev; futex_atomic_cmpxchg_inatomic() local
128 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic()
132 *uval = prev; futex_atomic_cmpxchg_inatomic()
/linux-4.4.14/drivers/infiniband/hw/mthca/
H A Dmthca_mcg.c54 * If GID is found in MGM or MGM is empty, *index = *hash, *prev = -1
57 * if GID is found in AMGM, *index = index in AMGM, *prev = index of
60 * If no AMGM exists for given gid, *index = -1, *prev = index of last
65 u16 *hash, int *prev, int *index) find_mgm()
89 *prev = -1; find_mgm()
109 *prev = *index; find_mgm()
126 int index, prev; mthca_multicast_attach() local
138 err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); mthca_multicast_attach()
191 err = mthca_READ_MGM(dev, prev, mailbox); mthca_multicast_attach()
199 err = mthca_WRITE_MGM(dev, prev, mailbox); mthca_multicast_attach()
220 int prev, index; mthca_multicast_detach() local
231 err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); mthca_multicast_detach()
266 if (prev == -1) { mthca_multicast_detach()
291 err = mthca_READ_MGM(dev, prev, mailbox); mthca_multicast_detach()
299 err = mthca_WRITE_MGM(dev, prev, mailbox); mthca_multicast_detach()
63 find_mgm(struct mthca_dev *dev, u8 *gid, struct mthca_mailbox *mgm_mailbox, u16 *hash, int *prev, int *index) find_mgm() argument
/linux-4.4.14/tools/perf/scripts/python/
H A Dsched-migration.py102 def sched_switch(self, prev, prev_state, next):
106 and prev in self.tasks:
110 event = RunqueueEventSleep(prev)
113 if prev in self.tasks:
115 next_tasks.remove(prev)
117 next_tasks.append(prev)
160 def __init__(self, start, prev):
162 self.prev = prev
166 if prev is not None:
167 self.total_load = prev.total_load
168 self.rqs = prev.rqs.copy()
177 def sched_switch(self, ts_list, prev, prev_state, next, cpu):
178 old_rq = self.prev.rqs[cpu]
179 new_rq = old_rq.sched_switch(prev, prev_state, next)
192 old_rq = self.prev.rqs[old_cpu]
197 new_rq = self.prev.rqs[new_cpu]
209 old_rq = self.prev.rqs[cpu]
339 print "Sched switch event rejected ts: %s cpu: %d prev: %s(%d) next: %s(%d)" % \
/linux-4.4.14/drivers/gpu/drm/nouveau/
H A Dnv17_fence.c34 struct nouveau_channel *prev, struct nouveau_channel *chan) nv17_fence_sync()
36 struct nouveau_cli *cli = (void *)prev->user.client; nv17_fence_sync()
50 ret = RING_SPACE(prev, 5); nv17_fence_sync()
52 BEGIN_NV04(prev, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 4); nv17_fence_sync()
53 OUT_RING (prev, fctx->sema.handle); nv17_fence_sync()
54 OUT_RING (prev, 0); nv17_fence_sync()
55 OUT_RING (prev, value + 0); nv17_fence_sync()
56 OUT_RING (prev, value + 1); nv17_fence_sync()
57 FIRE_RING (prev); nv17_fence_sync()
33 nv17_fence_sync(struct nouveau_fence *fence, struct nouveau_channel *prev, struct nouveau_channel *chan) nv17_fence_sync() argument
H A Dnouveau_bo.h53 struct nouveau_bo *prev; nouveau_bo_ref() local
57 prev = *pnvbo; nouveau_bo_ref()
60 if (prev) { nouveau_bo_ref()
61 struct ttm_buffer_object *bo = &prev->bo; nouveau_bo_ref()
H A Dnouveau_fence.c410 struct nouveau_channel *prev = NULL; nouveau_fence_sync() local
416 prev = rcu_dereference(f->channel); nouveau_fence_sync()
417 if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) nouveau_fence_sync()
432 struct nouveau_channel *prev = NULL; nouveau_fence_sync() local
441 prev = rcu_dereference(f->channel); nouveau_fence_sync()
442 if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) nouveau_fence_sync()
/linux-4.4.14/drivers/char/agp/
H A Dfrontend.c68 struct agp_memory *prev; agp_remove_from_pool() local
76 prev = temp->prev; agp_remove_from_pool()
78 if (prev != NULL) { agp_remove_from_pool()
79 prev->next = next; agp_remove_from_pool()
81 next->prev = prev; agp_remove_from_pool()
86 next->prev = NULL; agp_remove_from_pool()
201 struct agp_memory *prev; agp_insert_into_pool() local
203 prev = agp_fe.current_controller->pool; agp_insert_into_pool()
205 if (prev != NULL) { agp_insert_into_pool()
206 prev->prev = temp; agp_insert_into_pool()
207 temp->next = prev; agp_insert_into_pool()
232 struct agp_file_private *prev; agp_insert_file_private() local
234 prev = agp_fe.file_priv_list; agp_insert_file_private()
236 if (prev != NULL) agp_insert_file_private()
237 prev->prev = priv; agp_insert_file_private()
238 priv->next = prev; agp_insert_file_private()
245 struct agp_file_private *prev; agp_remove_file_private() local
248 prev = priv->prev; agp_remove_file_private()
250 if (prev != NULL) { agp_remove_file_private()
251 prev->next = next; agp_remove_file_private()
254 next->prev = prev; agp_remove_file_private()
258 next->prev = NULL; agp_remove_file_private()
328 prev_controller->prev = controller; agp_insert_controller()
377 prev_controller = controller->prev; agp_remove_controller()
383 next_controller->prev = prev_controller; agp_remove_controller()
387 next_controller->prev = NULL; agp_remove_controller()
507 prev_client->prev = client; agp_insert_client()
541 prev_client = client->prev; agp_remove_client()
547 next_client->prev = prev_client; agp_remove_client()
551 next_client->prev = NULL; agp_remove_client()
/linux-4.4.14/crypto/
H A Dxcbc.c53 * | prev (block size)
87 u8 *prev = PTR_ALIGN(&ctx->ctx[0], alignmask + 1) + bs; crypto_xcbc_digest_init() local
90 memset(prev, 0, bs); crypto_xcbc_digest_init()
105 u8 *prev = odds + bs; crypto_xcbc_digest_update() local
119 crypto_xor(prev, odds, bs); crypto_xcbc_digest_update()
120 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_xcbc_digest_update()
127 crypto_xor(prev, p, bs); crypto_xcbc_digest_update()
128 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_xcbc_digest_update()
152 u8 *prev = odds + bs; crypto_xcbc_digest_final() local
169 crypto_xor(prev, odds, bs); crypto_xcbc_digest_final()
170 crypto_xor(prev, consts + offset, bs); crypto_xcbc_digest_final()
172 crypto_cipher_encrypt_one(tfm, out, prev); crypto_xcbc_digest_final()
H A Dcmac.c46 * | prev (block size)
113 u8 *prev = PTR_ALIGN((void *)ctx->ctx, alignmask + 1) + bs; crypto_cmac_digest_init() local
116 memset(prev, 0, bs); crypto_cmac_digest_init()
131 u8 *prev = odds + bs; crypto_cmac_digest_update() local
145 crypto_xor(prev, odds, bs); crypto_cmac_digest_update()
146 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_cmac_digest_update()
153 crypto_xor(prev, p, bs); crypto_cmac_digest_update()
154 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_cmac_digest_update()
178 u8 *prev = odds + bs; crypto_cmac_digest_final() local
195 crypto_xor(prev, odds, bs); crypto_cmac_digest_final()
196 crypto_xor(prev, consts + offset, bs); crypto_cmac_digest_final()
198 crypto_cipher_encrypt_one(tfm, out, prev); crypto_cmac_digest_final()
/linux-4.4.14/sound/isa/gus/
H A Dgus_mem.c54 nblock->prev = pblock->prev; snd_gf1_mem_xalloc()
56 pblock->prev = nblock; snd_gf1_mem_xalloc()
60 nblock->prev->next = nblock; snd_gf1_mem_xalloc()
68 nblock->prev = NULL; snd_gf1_mem_xalloc()
71 nblock->prev = alloc->last; snd_gf1_mem_xalloc()
88 block->next->prev = NULL; snd_gf1_mem_xfree()
90 block->prev->next = block->next; snd_gf1_mem_xfree()
92 block->next->prev = block->prev; snd_gf1_mem_xfree()
95 alloc->last = block->prev; snd_gf1_mem_xfree()
96 if (block->prev) snd_gf1_mem_xfree()
97 block->prev->next = NULL; snd_gf1_mem_xfree()
99 block->next->prev = block->prev; snd_gf1_mem_xfree()
100 if (block->prev) snd_gf1_mem_xfree()
101 block->prev->next = block->next; snd_gf1_mem_xfree()
154 block->prev = block->next = NULL; snd_gf1_mem_find()
/linux-4.4.14/arch/sparc/lib/
H A Datomic32.c159 u32 prev; __cmpxchg_u32() local
162 if ((prev = *ptr) == old) __cmpxchg_u32()
166 return (unsigned long)prev; __cmpxchg_u32()
173 u32 prev; __xchg_u32() local
176 prev = *ptr; __xchg_u32()
180 return (unsigned long)prev; __xchg_u32()
H A Dmemcpy.S100 #define SMOVE_CHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \
106 or %t5, %prev, %t5; \
107 sll %t1, shil, %prev; \
112 or %t1, %prev, %t1; \
115 sll %t3, shil, %prev; \
118 #define SMOVE_ALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \
124 or %t4, %prev, %t0; \
125 sll %t1, shil, %prev; \
130 or %t4, %prev, %t2; \
131 sll %t3, shil, %prev; \
/linux-4.4.14/arch/powerpc/platforms/cell/spufs/
H A Dswitch.c1791 static int quiece_spu(struct spu_state *prev, struct spu *spu) quiece_spu() argument
1803 if (check_spu_isolate(prev, spu)) { /* Step 2. */ quiece_spu()
1806 disable_interrupts(prev, spu); /* Step 3. */ quiece_spu()
1807 set_watchdog_timer(prev, spu); /* Step 4. */ quiece_spu()
1808 inhibit_user_access(prev, spu); /* Step 5. */ quiece_spu()
1809 if (check_spu_isolate(prev, spu)) { /* Step 6. */ quiece_spu()
1812 set_switch_pending(prev, spu); /* Step 7. */ quiece_spu()
1813 save_mfc_cntl(prev, spu); /* Step 8. */ quiece_spu()
1814 save_spu_runcntl(prev, spu); /* Step 9. */ quiece_spu()
1815 save_mfc_sr1(prev, spu); /* Step 10. */ quiece_spu()
1816 save_spu_status(prev, spu); /* Step 11. */ quiece_spu()
1817 save_mfc_stopped_status(prev, spu); /* Step 12. */ quiece_spu()
1818 halt_mfc_decr(prev, spu); /* Step 13. */ quiece_spu()
1819 save_timebase(prev, spu); /* Step 14. */ quiece_spu()
1820 remove_other_spu_access(prev, spu); /* Step 15. */ quiece_spu()
1821 do_mfc_mssync(prev, spu); /* Step 16. */ quiece_spu()
1822 issue_mfc_tlbie(prev, spu); /* Step 17. */ quiece_spu()
1823 handle_pending_interrupts(prev, spu); /* Step 18. */ quiece_spu()
1828 static void save_csa(struct spu_state *prev, struct spu *spu) save_csa() argument
1835 save_mfc_queues(prev, spu); /* Step 19. */ save_csa()
1836 save_ppu_querymask(prev, spu); /* Step 20. */ save_csa()
1837 save_ppu_querytype(prev, spu); /* Step 21. */ save_csa()
1838 save_ppu_tagstatus(prev, spu); /* NEW. */ save_csa()
1839 save_mfc_csr_tsq(prev, spu); /* Step 22. */ save_csa()
1840 save_mfc_csr_cmd(prev, spu); /* Step 23. */ save_csa()
1841 save_mfc_csr_ato(prev, spu); /* Step 24. */ save_csa()
1842 save_mfc_tclass_id(prev, spu); /* Step 25. */ save_csa()
1843 set_mfc_tclass_id(prev, spu); /* Step 26. */ save_csa()
1844 save_mfc_cmd(prev, spu); /* Step 26a - moved from 44. */ save_csa()
1845 purge_mfc_queue(prev, spu); /* Step 27. */ save_csa()
1846 wait_purge_complete(prev, spu); /* Step 28. */ save_csa()
1847 setup_mfc_sr1(prev, spu); /* Step 30. */ save_csa()
1848 save_spu_npc(prev, spu); /* Step 31. */ save_csa()
1849 save_spu_privcntl(prev, spu); /* Step 32. */ save_csa()
1850 reset_spu_privcntl(prev, spu); /* Step 33. */ save_csa()
1851 save_spu_lslr(prev, spu); /* Step 34. */ save_csa()
1852 reset_spu_lslr(prev, spu); /* Step 35. */ save_csa()
1853 save_spu_cfg(prev, spu); /* Step 36. */ save_csa()
1854 save_pm_trace(prev, spu); /* Step 37. */ save_csa()
1855 save_mfc_rag(prev, spu); /* Step 38. */ save_csa()
1856 save_ppu_mb_stat(prev, spu); /* Step 39. */ save_csa()
1857 save_ppu_mb(prev, spu); /* Step 40. */ save_csa()
1858 save_ppuint_mb(prev, spu); /* Step 41. */ save_csa()
1859 save_ch_part1(prev, spu); /* Step 42. */ save_csa()
1860 save_spu_mb(prev, spu); /* Step 43. */ save_csa()
1861 reset_ch(prev, spu); /* Step 45. */ save_csa()
1864 static void save_lscsa(struct spu_state *prev, struct spu *spu) save_lscsa() argument
1872 resume_mfc_queue(prev, spu); /* Step 46. */ save_lscsa()
1874 setup_mfc_slbs(prev, spu, spu_save_code, sizeof(spu_save_code)); save_lscsa()
1875 set_switch_active(prev, spu); /* Step 48. */ save_lscsa()
1876 enable_interrupts(prev, spu); /* Step 49. */ save_lscsa()
1877 save_ls_16kb(prev, spu); /* Step 50. */ save_lscsa()
1878 set_spu_npc(prev, spu); /* Step 51. */ save_lscsa()
1879 set_signot1(prev, spu); /* Step 52. */ save_lscsa()
1880 set_signot2(prev, spu); /* Step 53. */ save_lscsa()
1881 send_save_code(prev, spu); /* Step 54. */ save_lscsa()
1882 set_ppu_querymask(prev, spu); /* Step 55. */ save_lscsa()
1883 wait_tag_complete(prev, spu); /* Step 56. */ save_lscsa()
1884 wait_spu_stopped(prev, spu); /* Step 57. */ save_lscsa()
1932 static void harvest(struct spu_state *prev, struct spu *spu) harvest() argument
1940 disable_interrupts(prev, spu); /* Step 2. */ harvest()
1941 inhibit_user_access(prev, spu); /* Step 3. */ harvest()
1942 terminate_spu_app(prev, spu); /* Step 4. */ harvest()
1943 set_switch_pending(prev, spu); /* Step 5. */ harvest()
1945 remove_other_spu_access(prev, spu); /* Step 6. */ harvest()
1946 suspend_mfc_and_halt_decr(prev, spu); /* Step 7. */ harvest()
1947 wait_suspend_mfc_complete(prev, spu); /* Step 8. */ harvest()
1948 if (!suspend_spe(prev, spu)) /* Step 9. */ harvest()
1949 clear_spu_status(prev, spu); /* Step 10. */ harvest()
1950 do_mfc_mssync(prev, spu); /* Step 11. */ harvest()
1951 issue_mfc_tlbie(prev, spu); /* Step 12. */ harvest()
1952 handle_pending_interrupts(prev, spu); /* Step 13. */ harvest()
1953 purge_mfc_queue(prev, spu); /* Step 14. */ harvest()
1954 wait_purge_complete(prev, spu); /* Step 15. */ harvest()
1955 reset_spu_privcntl(prev, spu); /* Step 16. */ harvest()
1956 reset_spu_lslr(prev, spu); /* Step 17. */ harvest()
1957 setup_mfc_sr1(prev, spu); /* Step 18. */ harvest()
1959 reset_ch_part1(prev, spu); /* Step 20. */ harvest()
1960 reset_ch_part2(prev, spu); /* Step 21. */ harvest()
1961 enable_interrupts(prev, spu); /* Step 22. */ harvest()
1962 set_switch_active(prev, spu); /* Step 23. */ harvest()
1963 set_mfc_tclass_id(prev, spu); /* Step 24. */ harvest()
1964 resume_mfc_queue(prev, spu); /* Step 25. */ harvest()
2038 static int __do_spu_save(struct spu_state *prev, struct spu *spu) __do_spu_save() argument
2054 rc = quiece_spu(prev, spu); /* Steps 2-16. */ __do_spu_save()
2059 harvest(prev, spu); __do_spu_save()
2065 save_csa(prev, spu); /* Steps 17-43. */ __do_spu_save()
2066 save_lscsa(prev, spu); /* Steps 44-53. */ __do_spu_save()
2067 return check_save_status(prev, spu); /* Step 54. */ __do_spu_save()
2103 * @prev: pointer to SPU context save area, to be saved.
2108 int spu_save(struct spu_state *prev, struct spu *spu) spu_save() argument
2113 rc = __do_spu_save(prev, spu); /* Steps 2-53. */ spu_save()
/linux-4.4.14/arch/s390/lib/
H A Dspinlock.c186 void _raw_write_lock_wait(arch_rwlock_t *rw, unsigned int prev) _raw_write_lock_wait() argument
202 prev = __RAW_LOCK(&rw->lock, 0x80000000, __RAW_OP_OR); _raw_write_lock_wait()
203 old = prev; _raw_write_lock_wait()
205 if ((old & 0x7fffffff) == 0 && (int) prev >= 0) _raw_write_lock_wait()
217 unsigned int owner, old, prev; _raw_write_lock_wait() local
220 prev = 0x80000000; _raw_write_lock_wait()
232 prev = old; _raw_write_lock_wait()
235 if ((old & 0x7fffffff) == 0 && (int) prev >= 0) _raw_write_lock_wait()
/linux-4.4.14/tools/perf/
H A Dperf-completion.sh73 prev)
74 prev=${words_[$cword_-1]}
162 elif [[ $prev == @("-e"|"--event") &&
232 local _ret=1 cur cword prev
234 prev=${words[CURRENT-1]}
248 local cur words cword prev
250 _get_comp_words_by_ref -n =: cur words cword prev
252 __perf_get_comp_words_by_ref -n =: cur words cword prev
H A Dbuiltin-timechart.c726 struct io_sample *prev; pid_begin_io_sample() local
737 prev = c->io_samples; pid_begin_io_sample()
739 if (prev && prev->start_time && !prev->end_time) { pid_begin_io_sample()
745 c->io_samples = prev->next; pid_begin_io_sample()
746 free(prev); pid_begin_io_sample()
770 struct io_sample *sample, *prev; pid_end_io_sample() local
794 prev = sample->next; pid_end_io_sample()
800 if (prev && sample->start_time < prev->end_time) { pid_end_io_sample()
801 if (prev->err) /* try to make errors more visible */ pid_end_io_sample()
802 sample->start_time = prev->end_time; pid_end_io_sample()
804 prev->end_time = sample->start_time; pid_end_io_sample()
821 if (prev && pid_end_io_sample()
822 prev->type == sample->type && pid_end_io_sample()
823 prev->err == sample->err && pid_end_io_sample()
824 prev->fd == sample->fd && pid_end_io_sample()
825 prev->end_time + tchart->merge_dist >= sample->start_time) { pid_end_io_sample()
827 sample->bytes += prev->bytes; pid_end_io_sample()
828 sample->merges += prev->merges + 1; pid_end_io_sample()
830 sample->start_time = prev->start_time; pid_end_io_sample()
831 sample->next = prev->next; pid_end_io_sample()
832 free(prev); pid_end_io_sample()
968 struct per_pid *new_list, *p, *cursor, *prev; sort_pids() local
983 prev = NULL; sort_pids()
989 if (prev) { sort_pids()
990 p->next = prev->next; sort_pids()
991 prev->next = p; sort_pids()
1002 prev = cursor; sort_pids()
1005 prev->next = p; sort_pids()
/linux-4.4.14/include/uapi/linux/netfilter/
H A Dxt_limit.h18 unsigned long prev; /* moved to xt_limit_priv */ member in struct:xt_rateinfo
/linux-4.4.14/arch/um/include/asm/
H A Dmmu_context.h50 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
55 if(prev != next){ switch_mm()
56 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
/linux-4.4.14/kernel/sched/
H A Didle_task.c27 pick_next_task_idle(struct rq *rq, struct task_struct *prev) pick_next_task_idle() argument
29 put_prev_task(rq, prev); pick_next_task_idle()
48 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) put_prev_task_idle() argument
H A Dstop_task.c27 pick_next_task_stop(struct rq *rq, struct task_struct *prev) pick_next_task_stop() argument
34 put_prev_task(rq, prev); pick_next_task_stop()
58 static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) put_prev_task_stop() argument
H A Dcputime.c392 void vtime_common_task_switch(struct task_struct *prev) vtime_common_task_switch() argument
394 if (is_idle_task(prev)) vtime_common_task_switch()
395 vtime_account_idle(prev); vtime_common_task_switch()
397 vtime_account_system(prev); vtime_common_task_switch()
400 vtime_account_user(prev); vtime_common_task_switch()
402 arch_vtime_task_switch(prev); vtime_common_task_switch()
579 struct prev_cputime *prev, cputime_adjust()
586 raw_spin_lock_irqsave(&prev->lock, flags); cputime_adjust()
597 if (prev->stime + prev->utime >= rtime) cputime_adjust()
625 if (stime < prev->stime) cputime_adjust()
626 stime = prev->stime; cputime_adjust()
633 if (utime < prev->utime) { cputime_adjust()
634 utime = prev->utime; cputime_adjust()
639 prev->stime = stime; cputime_adjust()
640 prev->utime = utime; cputime_adjust()
642 *ut = prev->utime; cputime_adjust()
643 *st = prev->stime; cputime_adjust()
644 raw_spin_unlock_irqrestore(&prev->lock, flags); cputime_adjust()
764 void arch_vtime_task_switch(struct task_struct *prev) arch_vtime_task_switch() argument
766 write_seqlock(&prev->vtime_seqlock); arch_vtime_task_switch()
767 prev->vtime_snap_whence = VTIME_SLEEPING; arch_vtime_task_switch()
768 write_sequnlock(&prev->vtime_seqlock); arch_vtime_task_switch()
578 cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, cputime_t *ut, cputime_t *st) cputime_adjust() argument
H A Dstats.h128 * the idle task.) We are only called when prev != next.
132 struct task_struct *prev, struct task_struct *next) __sched_info_switch()
135 * prev now departs the cpu. It's not interesting to record __sched_info_switch()
139 if (prev != rq->idle) __sched_info_switch()
140 sched_info_depart(rq, prev); __sched_info_switch()
147 struct task_struct *prev, struct task_struct *next) sched_info_switch()
150 __sched_info_switch(rq, prev, next); sched_info_switch()
131 __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) __sched_info_switch() argument
146 sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) sched_info_switch() argument
H A Dauto_group.c128 struct autogroup *prev; autogroup_move_group() local
134 prev = p->signal->autogroup; autogroup_move_group()
135 if (prev == ag) { autogroup_move_group()
149 autogroup_kref_put(prev); autogroup_move_group()
/linux-4.4.14/arch/openrisc/include/asm/
H A Dmmu_context.h26 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next,
31 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
/linux-4.4.14/drivers/isdn/divert/
H A Disdn_divert.c30 struct call_struc *prev; member in struct:call_struc
38 struct deflect_struc *next, *prev; member in struct:deflect_struc
93 if (cs->prev) deflect_timer_expire()
94 cs->prev->next = cs->next; /* forward link */ deflect_timer_expire()
98 cs->next->prev = cs->prev; /* back link */ deflect_timer_expire()
191 cs->prev = NULL; cf_command()
295 ds->prev = table_tail; /* previous entry */ insertrule()
297 if (ds->prev) insertrule()
298 ds->prev->next = ds; /* last forward */ insertrule()
304 ds->prev = ds1->prev; /* prev entry */ insertrule()
305 ds1->prev = ds; /* backward chain old element */ insertrule()
306 if (!ds->prev) insertrule()
350 ds->next->prev = ds->prev; /* backward chain */ deleterule()
352 table_tail = ds->prev; /* end of chain */ deleterule()
354 if (ds->prev) deleterule()
355 ds->prev->next = ds->next; /* forward chain */ deleterule()
474 cs->prev = NULL; isdn_divert_icall()
518 cs->prev = NULL; isdn_divert_icall()
756 if (cs1->prev) prot_stat_callback()
757 cs1->prev->next = cs1->next; /* forward link */ prot_stat_callback()
761 cs1->next->prev = cs1->prev; /* back link */ prot_stat_callback()
812 if (cs1->prev) isdn_divert_stat_callback()
813 cs1->prev->next = cs1->next; /* forward link */ isdn_divert_stat_callback()
817 cs1->next->prev = cs1->prev; /* back link */ isdn_divert_stat_callback()
/linux-4.4.14/net/ceph/
H A Dpagelist.c11 struct page *page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_unmap_tail()
119 c->page_lru = pl->head.prev; ceph_pagelist_set_cursor()
137 while (pl->head.prev != c->page_lru) { ceph_pagelist_truncate()
138 page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_truncate()
145 page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_truncate()
/linux-4.4.14/net/netfilter/
H A Dxt_limit.c21 unsigned long prev; member in struct:xt_limit_priv
43 `prev' tracks the last packet hit: you gain one credit per jiffy.
76 priv->credit += (now - xchg(&priv->prev, now)) * CREDITS_PER_JIFFY; limit_mt()
123 priv->prev = jiffies; limit_mt_check()
144 compat_ulong_t prev; member in struct:compat_xt_rateinfo
151 /* To keep the full "prev" timestamp, the upper 32 bits are stored in the
159 .prev = cm->prev | (unsigned long)cm->master << 32, limit_mt_compat_from_user()
173 .prev = m->prev, limit_mt_compat_to_user()
177 .master = m->prev >> 32, limit_mt_compat_to_user()
/linux-4.4.14/net/netfilter/ipset/
H A Dip_set_list_set.c197 struct set_elem *e, *next, *prev = NULL; list_set_utest() local
205 prev = e; list_set_utest()
216 ret = prev && prev->id == d->refid; list_set_utest()
244 struct set_elem *e, *n, *prev, *next; list_set_uadd() local
251 n = prev = next = NULL; list_set_uadd()
263 prev = e; list_set_uadd()
268 (d->before < 0 && !prev)) list_set_uadd()
290 /* Insert before prev element */ list_set_uadd()
291 if (prev->list.prev != &map->members) list_set_uadd()
292 n = list_prev_entry(prev, list); list_set_uadd()
310 else if (prev) list_set_uadd()
311 list_add_rcu(&e->list, &prev->list); list_set_uadd()
324 struct set_elem *e, *next, *prev = NULL; list_set_udel() local
331 prev = e; list_set_udel()
341 if (!prev || prev->id != d->refid) list_set_udel()
/linux-4.4.14/arch/powerpc/boot/
H A Dops.h47 void *(*find_node_by_prop_value)(const void *prev,
50 void *(*find_node_by_compatible)(const void *prev,
146 static inline void *find_node_by_prop_value(const void *prev, find_node_by_prop_value() argument
151 return dt_ops.find_node_by_prop_value(prev, propname, find_node_by_prop_value()
157 static inline void *find_node_by_prop_value_str(const void *prev, find_node_by_prop_value_str() argument
161 return find_node_by_prop_value(prev, propname, propval, find_node_by_prop_value_str()
165 static inline void *find_node_by_devtype(const void *prev, find_node_by_devtype() argument
168 return find_node_by_prop_value_str(prev, "device_type", type); find_node_by_devtype()
184 static inline void *find_node_by_compatible(const void *prev, find_node_by_compatible() argument
188 return dt_ops.find_node_by_compatible(prev, compat); find_node_by_compatible()
H A Dlibfdt-wrapper.c131 static void *fdt_wrapper_find_node_by_prop_value(const void *prev, fdt_wrapper_find_node_by_prop_value() argument
136 int offset = fdt_node_offset_by_prop_value(fdt, devp_offset_find(prev), fdt_wrapper_find_node_by_prop_value()
141 static void *fdt_wrapper_find_node_by_compatible(const void *prev, fdt_wrapper_find_node_by_compatible() argument
144 int offset = fdt_node_offset_by_compatible(fdt, devp_offset_find(prev), fdt_wrapper_find_node_by_compatible()
/linux-4.4.14/arch/x86/kernel/cpu/
H A Dperf_event_msr.c154 u64 prev, now; msr_event_update() local
159 prev = local64_read(&event->hw.prev_count); msr_event_update()
162 if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev) msr_event_update()
165 delta = now - prev; msr_event_update()
169 local64_add(now - prev, &event->count); msr_event_update()
/linux-4.4.14/sound/synth/
H A Dutil_mem.c101 return __snd_util_memblk_new(hdr, units, p->prev); __snd_util_mem_alloc()
107 * the block is linked next to prev
111 struct list_head *prev) __snd_util_memblk_new()
120 if (prev == &hdr->block) __snd_util_memblk_new()
123 struct snd_util_memblk *p = get_memblk(prev); __snd_util_memblk_new()
127 list_add(&blk->list, prev); __snd_util_memblk_new()
110 __snd_util_memblk_new(struct snd_util_memhdr *hdr, unsigned int units, struct list_head *prev) __snd_util_memblk_new() argument
/linux-4.4.14/drivers/gpu/drm/
H A Ddrm_lock.c180 unsigned int old, new, prev; drm_lock_take() local
193 prev = cmpxchg(lock, old, new); drm_lock_take()
194 } while (prev != old); drm_lock_take()
229 unsigned int old, new, prev; drm_lock_transfer() local
236 prev = cmpxchg(lock, old, new); drm_lock_transfer()
237 } while (prev != old); drm_lock_transfer()
254 unsigned int old, new, prev; drm_legacy_lock_free() local
269 prev = cmpxchg(lock, old, new); drm_legacy_lock_free()
270 } while (prev != old); drm_legacy_lock_free()
315 unsigned int old, prev; drm_legacy_idlelock_release() local
323 prev = cmpxchg(lock, old, DRM_KERNEL_CONTEXT); drm_legacy_idlelock_release()
324 } while (prev != old); drm_legacy_idlelock_release()
/linux-4.4.14/net/bridge/netfilter/
H A Debt_limit.c41 info->credit += (now - xchg(&info->prev, now)) * CREDITS_PER_JIFFY; ebt_limit_mt()
81 info->prev = jiffies; ebt_limit_mt_check()
96 compat_ulong_t prev; member in struct:ebt_compat_limit_info
/linux-4.4.14/arch/metag/kernel/
H A Dprocess.c301 struct task_struct *__sched __switch_to(struct task_struct *prev, __switch_to() argument
307 to.Switch.pPara = prev; __switch_to()
310 if (prev->thread.user_flags & TBICTX_FPAC_BIT) { __switch_to()
311 struct pt_regs *regs = task_pt_regs(prev); __switch_to()
314 state.Sig.SaveMask = prev->thread.user_flags; __switch_to()
317 if (!prev->thread.fpu_context) __switch_to()
318 alloc_fpu_context(&prev->thread); __switch_to()
319 if (prev->thread.fpu_context) __switch_to()
320 __TBICtxFPUSave(state, prev->thread.fpu_context); __switch_to()
326 if (prev->thread.fpu_context) __switch_to()
327 prev->thread.fpu_context->needs_restore = true; __switch_to()
331 from = __TBISwitch(to, &prev->thread.kernel_context); __switch_to()
/linux-4.4.14/drivers/usb/image/
H A Dmicrotek.h31 struct mts_desc *prev; member in struct:mts_desc
/linux-4.4.14/drivers/media/pci/zoran/
H A Dvideocodec.c172 struct attached_list *a, *prev; videocodec_detach() local
192 prev = NULL; videocodec_detach()
208 if (prev == NULL) { videocodec_detach()
213 prev->next = a->next; videocodec_detach()
223 prev = a; videocodec_detach()
271 struct codec_list *prev = NULL, *h = codeclist_top; videocodec_unregister() local
300 if (prev == NULL) { videocodec_unregister()
305 prev->next = h->next; videocodec_unregister()
312 prev = h; videocodec_unregister()
/linux-4.4.14/drivers/input/joystick/
H A Djoydump.c52 struct joydump *dump, *prev; /* one entry each */ joydump_connect() local
123 prev = dump; joydump_connect()
132 for (i = 1; i < t; i++, dump++, prev++) { joydump_connect()
134 i, dump->time - prev->time); joydump_connect()
/linux-4.4.14/drivers/net/ethernet/sfc/
H A Dmdio_10g.c235 struct ethtool_cmd prev = { .cmd = ETHTOOL_GSET }; efx_mdio_set_settings() local
237 efx->phy_op->get_settings(efx, &prev); efx_mdio_set_settings()
239 if (ecmd->advertising == prev.advertising && efx_mdio_set_settings()
240 ethtool_cmd_speed(ecmd) == ethtool_cmd_speed(&prev) && efx_mdio_set_settings()
241 ecmd->duplex == prev.duplex && efx_mdio_set_settings()
242 ecmd->port == prev.port && efx_mdio_set_settings()
243 ecmd->autoneg == prev.autoneg) efx_mdio_set_settings()
247 if (prev.port != PORT_TP || ecmd->port != PORT_TP) efx_mdio_set_settings()
252 (ecmd->advertising | SUPPORTED_Autoneg) & ~prev.supported) efx_mdio_set_settings()
/linux-4.4.14/arch/arm/mach-omap2/
H A Dsdrc2xxx.c85 u32 prev = curr_perf_level; omap2xxx_sdrc_reprogram() local
89 return prev; omap2xxx_sdrc_reprogram()
96 return prev; omap2xxx_sdrc_reprogram()
113 return prev; omap2xxx_sdrc_reprogram()
/linux-4.4.14/drivers/acpi/
H A Dnfit.c233 struct nfit_table_prev *prev, add_spa()
240 list_for_each_entry(nfit_spa, &prev->spas, list) { add_spa()
260 struct nfit_table_prev *prev, add_memdev()
267 list_for_each_entry(nfit_memdev, &prev->memdevs, list) add_memdev()
286 struct nfit_table_prev *prev, add_dcr()
293 list_for_each_entry(nfit_dcr, &prev->dcrs, list) add_dcr()
311 struct nfit_table_prev *prev, add_bdw()
318 list_for_each_entry(nfit_bdw, &prev->bdws, list) add_bdw()
336 struct nfit_table_prev *prev, add_idt()
343 list_for_each_entry(nfit_idt, &prev->idts, list) add_idt()
361 struct nfit_table_prev *prev, add_flush()
368 list_for_each_entry(nfit_flush, &prev->flushes, list) add_flush()
386 struct nfit_table_prev *prev, void *table, const void *end) add_table()
404 if (!add_spa(acpi_desc, prev, table)) add_table()
408 if (!add_memdev(acpi_desc, prev, table)) add_table()
412 if (!add_dcr(acpi_desc, prev, table)) add_table()
416 if (!add_bdw(acpi_desc, prev, table)) add_table()
420 if (!add_idt(acpi_desc, prev, table)) add_table()
424 if (!add_flush(acpi_desc, prev, table)) add_table()
1612 struct nfit_table_prev *prev) acpi_nfit_check_deletions()
1616 if (!list_empty(&prev->spas) || acpi_nfit_check_deletions()
1617 !list_empty(&prev->memdevs) || acpi_nfit_check_deletions()
1618 !list_empty(&prev->dcrs) || acpi_nfit_check_deletions()
1619 !list_empty(&prev->bdws) || acpi_nfit_check_deletions()
1620 !list_empty(&prev->idts) || acpi_nfit_check_deletions()
1621 !list_empty(&prev->flushes)) { acpi_nfit_check_deletions()
1631 struct nfit_table_prev prev; acpi_nfit_init() local
1638 INIT_LIST_HEAD(&prev.spas); acpi_nfit_init()
1639 INIT_LIST_HEAD(&prev.memdevs); acpi_nfit_init()
1640 INIT_LIST_HEAD(&prev.dcrs); acpi_nfit_init()
1641 INIT_LIST_HEAD(&prev.bdws); acpi_nfit_init()
1642 INIT_LIST_HEAD(&prev.idts); acpi_nfit_init()
1643 INIT_LIST_HEAD(&prev.flushes); acpi_nfit_init()
1645 list_cut_position(&prev.spas, &acpi_desc->spas, acpi_nfit_init()
1646 acpi_desc->spas.prev); acpi_nfit_init()
1647 list_cut_position(&prev.memdevs, &acpi_desc->memdevs, acpi_nfit_init()
1648 acpi_desc->memdevs.prev); acpi_nfit_init()
1649 list_cut_position(&prev.dcrs, &acpi_desc->dcrs, acpi_nfit_init()
1650 acpi_desc->dcrs.prev); acpi_nfit_init()
1651 list_cut_position(&prev.bdws, &acpi_desc->bdws, acpi_nfit_init()
1652 acpi_desc->bdws.prev); acpi_nfit_init()
1653 list_cut_position(&prev.idts, &acpi_desc->idts, acpi_nfit_init()
1654 acpi_desc->idts.prev); acpi_nfit_init()
1655 list_cut_position(&prev.flushes, &acpi_desc->flushes, acpi_nfit_init()
1656 acpi_desc->flushes.prev); acpi_nfit_init()
1661 data = add_table(acpi_desc, &prev, data, end); acpi_nfit_init()
1670 rc = acpi_nfit_check_deletions(acpi_desc, &prev); acpi_nfit_init()
232 add_spa(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_system_address *spa) add_spa() argument
259 add_memdev(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_memory_map *memdev) add_memdev() argument
285 add_dcr(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_control_region *dcr) add_dcr() argument
310 add_bdw(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_data_region *bdw) add_bdw() argument
335 add_idt(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_interleave *idt) add_idt() argument
360 add_flush(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_flush_address *flush) add_flush() argument
385 add_table(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, void *table, const void *end) add_table() argument
1611 acpi_nfit_check_deletions(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev) acpi_nfit_check_deletions() argument
/linux-4.4.14/fs/nfs/
H A Dpagelist.c269 * @prev - the previous request in page group, or NULL if @req is the first
273 nfs_page_group_init(struct nfs_page *req, struct nfs_page *prev) nfs_page_group_init() argument
276 WARN_ON_ONCE(prev == req); nfs_page_group_init()
278 if (!prev) { nfs_page_group_init()
284 WARN_ON_ONCE(prev->wb_this_page != prev->wb_head); nfs_page_group_init()
285 WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &prev->wb_head->wb_flags)); nfs_page_group_init()
286 req->wb_head = prev->wb_head; nfs_page_group_init()
287 req->wb_this_page = prev->wb_this_page; nfs_page_group_init()
288 prev->wb_this_page = req; nfs_page_group_init()
297 if (test_bit(PG_INODE_REF, &prev->wb_head->wb_flags)) { nfs_page_group_init()
488 * @prev: previous request in desc, or NULL
495 struct nfs_page *prev, struct nfs_page *req) nfs_generic_pg_test()
921 * @prev: pointer to nfs_page
924 * The nfs_page structures 'prev' and 'req' are compared to ensure that the
930 static bool nfs_can_coalesce_requests(struct nfs_page *prev, nfs_can_coalesce_requests() argument
937 if (prev) { nfs_can_coalesce_requests()
938 if (!nfs_match_open_context(req->wb_context, prev->wb_context)) nfs_can_coalesce_requests()
945 prev->wb_lock_context)) nfs_can_coalesce_requests()
947 if (req_offset(req) != req_offset(prev) + prev->wb_bytes) nfs_can_coalesce_requests()
949 if (req->wb_page == prev->wb_page) { nfs_can_coalesce_requests()
950 if (req->wb_pgbase != prev->wb_pgbase + prev->wb_bytes) nfs_can_coalesce_requests()
954 prev->wb_pgbase + prev->wb_bytes != PAGE_CACHE_SIZE) nfs_can_coalesce_requests()
958 size = pgio->pg_ops->pg_test(pgio, prev, req); nfs_can_coalesce_requests()
978 struct nfs_page *prev = NULL; nfs_pageio_do_add_request() local
981 prev = nfs_list_entry(mirror->pg_list.prev); nfs_pageio_do_add_request()
987 if (!nfs_can_coalesce_requests(prev, req, desc)) nfs_pageio_do_add_request()
1271 struct nfs_page *prev; nfs_pageio_cond_complete() local
1277 prev = nfs_list_entry(mirror->pg_list.prev); nfs_pageio_cond_complete()
1278 if (index != prev->wb_index + 1) nfs_pageio_cond_complete()
494 nfs_generic_pg_test(struct nfs_pageio_descriptor *desc, struct nfs_page *prev, struct nfs_page *req) nfs_generic_pg_test() argument
/linux-4.4.14/fs/hfs/
H A Dbnode.c144 be32_to_cpu(desc.next), be32_to_cpu(desc.prev), hfs_bnode_dump()
179 if (node->prev) { hfs_bnode_unlink()
180 tmp = hfs_bnode_find(tree, node->prev); hfs_bnode_unlink()
194 tmp->prev = node->prev; hfs_bnode_unlink()
195 cnid = cpu_to_be32(tmp->prev); hfs_bnode_unlink()
196 hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, prev), 4); hfs_bnode_unlink()
199 tree->leaf_tail = node->prev; hfs_bnode_unlink()
202 if (!node->prev && !node->next) { hfs_bnode_unlink()
342 node->prev = be32_to_cpu(desc->prev); hfs_bnode_find()
/linux-4.4.14/sound/core/seq/
H A Dseq_prioq.c149 struct snd_seq_event_cell *cur, *prev; snd_seq_prioq_cell_in() local
179 prev = NULL; /* previous cell */ snd_seq_prioq_cell_in()
194 prev = cur; snd_seq_prioq_cell_in()
204 if (prev != NULL) snd_seq_prioq_cell_in()
205 prev->next = cell; snd_seq_prioq_cell_in()
294 struct snd_seq_event_cell *prev = NULL; snd_seq_prioq_leave() local
307 prev->next = cell->next; snd_seq_prioq_leave()
329 prev = cell; snd_seq_prioq_leave()
404 struct snd_seq_event_cell *prev = NULL; snd_seq_prioq_remove_events() local
420 prev->next = cell->next; snd_seq_prioq_remove_events()
437 prev = cell; snd_seq_prioq_remove_events()
/linux-4.4.14/fs/notify/dnotify/
H A Ddnotify.c93 struct dnotify_struct **prev; dnotify_handle_event() local
106 prev = &dn_mark->dn; dnotify_handle_event()
107 while ((dn = *prev) != NULL) { dnotify_handle_event()
109 prev = &dn->dn_next; dnotify_handle_event()
115 prev = &dn->dn_next; dnotify_handle_event()
117 *prev = dn->dn_next; dnotify_handle_event()
155 struct dnotify_struct **prev; dnotify_flush() local
171 prev = &dn_mark->dn; dnotify_flush()
172 while ((dn = *prev) != NULL) { dnotify_flush()
174 *prev = dn->dn_next; dnotify_flush()
179 prev = &dn->dn_next; dnotify_flush()
/linux-4.4.14/drivers/misc/mic/scif/
H A Dscif_rma_list.c31 struct scif_window *prev = list_entry(head, struct scif_window, list); scif_insert_tcw() local
37 curr = list_entry(head->prev, struct scif_window, list); scif_insert_tcw()
47 prev = curr; list_for_each()
49 list_add(&window->list, &prev->list);
60 struct scif_window *curr = NULL, *prev = NULL; scif_insert_window() local
68 prev = curr; list_for_each()
70 if (!prev)
73 list_add(&window->list, &prev->list);
/linux-4.4.14/drivers/mtd/
H A Dredboot.c180 struct fis_list *new_fl, **prev; parse_redboot_partitions() local
207 prev = &fl; parse_redboot_partitions()
208 while(*prev && (*prev)->img->flash_base < new_fl->img->flash_base) parse_redboot_partitions()
209 prev = &(*prev)->next; parse_redboot_partitions()
210 new_fl->next = *prev; parse_redboot_partitions()
211 *prev = new_fl; parse_redboot_partitions()
/linux-4.4.14/drivers/dma/
H A Dtxx9dmac.c174 return list_entry(dc->active_list.prev, txx9dmac_last_active()
186 desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); txx9dmac_last_child()
435 struct txx9dmac_desc *prev = NULL; txx9dmac_dequeue() local
440 if (prev) { txx9dmac_dequeue()
441 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_dequeue()
443 prev->txd.phys, ddev->descsize, txx9dmac_dequeue()
446 prev = txx9dmac_last_child(desc); txx9dmac_dequeue()
723 struct txx9dmac_desc *prev; txx9dmac_prep_dma_memcpy() local
735 prev = first = NULL; txx9dmac_prep_dma_memcpy()
785 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_prep_dma_memcpy()
787 prev->txd.phys, ddev->descsize, txx9dmac_prep_dma_memcpy()
791 prev = desc; txx9dmac_prep_dma_memcpy()
796 txx9dmac_desc_set_INTENT(ddev, prev); txx9dmac_prep_dma_memcpy()
798 desc_write_CHAR(dc, prev, 0); txx9dmac_prep_dma_memcpy()
800 prev->txd.phys, ddev->descsize, txx9dmac_prep_dma_memcpy()
817 struct txx9dmac_desc *prev; txx9dmac_prep_slave_sg() local
832 prev = first = NULL; txx9dmac_prep_slave_sg()
879 desc_write_CHAR(dc, prev, desc->txd.phys); for_each_sg()
881 prev->txd.phys, for_each_sg()
886 prev = desc; for_each_sg()
891 txx9dmac_desc_set_INTENT(ddev, prev);
893 desc_write_CHAR(dc, prev, 0);
895 prev->txd.phys, ddev->descsize,
947 struct txx9dmac_desc *prev) txx9dmac_chain_dynamic()
953 prev = txx9dmac_last_child(prev); txx9dmac_chain_dynamic()
956 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_chain_dynamic()
958 prev->txd.phys, ddev->descsize, txx9dmac_chain_dynamic()
962 channel_read_CHAR(dc) == prev->txd.phys) txx9dmac_chain_dynamic()
981 struct txx9dmac_desc *prev = txx9dmac_last_active(dc); txx9dmac_issue_pending() local
983 if (!(prev->txd.flags & DMA_PREP_INTERRUPT) || txx9dmac_issue_pending()
985 txx9dmac_chain_dynamic(dc, prev); txx9dmac_issue_pending()
946 txx9dmac_chain_dynamic(struct txx9dmac_chan *dc, struct txx9dmac_desc *prev) txx9dmac_chain_dynamic() argument
H A Dimg-mdc-dma.c294 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_dma_memcpy() local
313 if (prev) { mdc_prep_dma_memcpy()
314 prev->node_addr = curr_phys; mdc_prep_dma_memcpy()
315 prev->next_desc = curr; mdc_prep_dma_memcpy()
326 prev = curr; mdc_prep_dma_memcpy()
377 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_dma_cyclic() local
409 if (!prev) { mdc_prep_dma_cyclic()
413 prev->node_addr = curr_phys; mdc_prep_dma_cyclic()
414 prev->next_desc = curr; mdc_prep_dma_cyclic()
432 prev = curr; mdc_prep_dma_cyclic()
441 prev->node_addr = mdesc->list_phys; mdc_prep_dma_cyclic()
460 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_slave_sg() local
490 if (!prev) { for_each_sg()
494 prev->node_addr = curr_phys; for_each_sg()
495 prev->next_desc = curr; for_each_sg()
511 prev = curr; for_each_sg()
/linux-4.4.14/arch/arc/kernel/
H A Dfpu.c33 void fpu_save_restore(struct task_struct *prev, struct task_struct *next) fpu_save_restore() argument
35 unsigned int *saveto = &prev->thread.fpu.aux_dpfp[0].l; fpu_save_restore()
H A Dctx_sw.c26 unsigned int prev = (unsigned int)prev_task; __switch_to() local
103 /* last (ret value) = prev : although for ARC it mov r0, r0 */ __switch_to()
109 : "n"(KSP_WORD_OFF), "r"(next), "r"(prev) __switch_to()
/linux-4.4.14/arch/c6x/kernel/
H A Dswitch_to.S16 * void __switch_to(struct thread_info *prev,
22 || MV .L2X A4,B5 ; prev
/linux-4.4.14/kernel/trace/
H A Dtrace_sched_switch.c20 struct task_struct *prev, struct task_struct *next) probe_sched_switch()
25 tracing_record_cmdline(prev); probe_sched_switch()
19 probe_sched_switch(void *ignore, bool preempt, struct task_struct *prev, struct task_struct *next) probe_sched_switch() argument
H A Dtrace_stat.h16 void *(*stat_next)(void *prev, int idx);
/linux-4.4.14/net/ieee802154/6lowpan/
H A Dreassembly.c38 struct sk_buff *prev, struct net_device *ldev);
132 struct sk_buff *prev, *next; lowpan_frag_queue() local
165 prev = fq->q.fragments_tail; lowpan_frag_queue()
166 if (!prev || lowpan_frag_queue()
167 lowpan_802154_cb(prev)->d_offset < lowpan_frag_queue()
172 prev = NULL; lowpan_frag_queue()
177 prev = next; lowpan_frag_queue()
185 if (prev) lowpan_frag_queue()
186 prev->next = skb; lowpan_frag_queue()
207 res = lowpan_frag_reasm(fq, prev, ldev); lowpan_frag_queue()
226 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *prev, lowpan_frag_reasm() argument
235 if (prev) { lowpan_frag_reasm()
236 head = prev->next; lowpan_frag_reasm()
245 prev->next = fp; lowpan_frag_reasm()
/linux-4.4.14/arch/sparc/kernel/
H A Dsun4m_smp.c123 int *prev; smp4m_smp_done() local
127 prev = &first; for_each_online_cpu()
129 *prev = i; for_each_online_cpu()
130 prev = &cpu_data(i).next; for_each_online_cpu()
132 *prev = first;
/linux-4.4.14/drivers/md/
H A Ddm-cache-policy-smq.c39 unsigned prev:28; member in struct:entry
137 return to_entry(es, e->prev); l_prev()
150 e->prev = INDEXER_NULL; l_add_head()
153 head->prev = l->head = to_index(es, e); l_add_head()
166 e->prev = l->tail; l_add_tail()
180 struct entry *prev = l_prev(es, old); l_add_before() local
182 if (!prev) l_add_before()
186 e->prev = old->prev; l_add_before()
188 prev->next = old->prev = to_index(es, e); l_add_before()
197 struct entry *prev = l_prev(es, e); l_del() local
200 if (prev) l_del()
201 prev->next = e->next; l_del()
206 next->prev = e->prev; l_del()
208 l->tail = e->prev; l_del()
609 struct entry **prev) __h_lookup()
613 *prev = NULL; __h_lookup()
618 *prev = e; __h_lookup()
625 struct entry *e, struct entry *prev) __h_unlink()
627 if (prev) __h_unlink()
628 prev->hash_next = e->hash_next; __h_unlink()
638 struct entry *e, *prev; h_lookup() local
641 e = __h_lookup(ht, h, oblock, &prev); h_lookup()
642 if (e && prev) { h_lookup()
647 __h_unlink(ht, h, e, prev); h_lookup()
657 struct entry *prev; h_remove() local
663 e = __h_lookup(ht, h, e->oblock, &prev); h_remove()
665 __h_unlink(ht, h, e, prev); h_remove()
700 e->prev = INDEXER_NULL; init_entry()
608 __h_lookup(struct hash_table *ht, unsigned h, dm_oblock_t oblock, struct entry **prev) __h_lookup() argument
624 __h_unlink(struct hash_table *ht, unsigned h, struct entry *e, struct entry *prev) __h_unlink() argument
/linux-4.4.14/fs/gfs2/
H A Dxattr.c77 struct gfs2_ea_header *prev, void *private);
82 struct gfs2_ea_header *ea, *prev = NULL; ea_foreach_i() local
88 for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { ea_foreach_i()
97 error = ea_call(ip, bh, ea, prev, data); ea_foreach_i()
167 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_find_i()
182 el->el_prev = prev; ea_find_i()
215 * @prev:
227 struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed()
297 if (prev && !leave) { ea_dealloc_unstuffed()
300 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_dealloc_unstuffed()
301 prev->ea_rec_len = cpu_to_be32(len); ea_dealloc_unstuffed()
304 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_dealloc_unstuffed()
327 struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed()
339 error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); ea_remove_unstuffed()
366 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_list_i()
838 struct gfs2_ea_header *prev = el->el_prev; ea_set_remove_stuffed() local
843 if (!prev || !GFS2_EA_IS_STUFFED(ea)) { ea_set_remove_stuffed()
846 } else if (GFS2_EA2NEXT(prev) != ea) { ea_set_remove_stuffed()
847 prev = GFS2_EA2NEXT(prev); ea_set_remove_stuffed()
848 gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea); ea_set_remove_stuffed()
851 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_set_remove_stuffed()
852 prev->ea_rec_len = cpu_to_be32(len); ea_set_remove_stuffed()
855 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_set_remove_stuffed()
924 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_set_simple()
939 error = ea_remove_unstuffed(ip, bh, ea, prev, 1); ea_set_simple()
1091 struct gfs2_ea_header *prev = el->el_prev; ea_remove_stuffed() local
1101 if (prev) { ea_remove_stuffed()
1104 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_remove_stuffed()
1105 prev->ea_rec_len = cpu_to_be32(len); ea_remove_stuffed()
1108 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_remove_stuffed()
166 ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_find_i() argument
225 ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() argument
325 ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() argument
365 ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_list_i() argument
923 ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_set_simple() argument
/linux-4.4.14/drivers/usb/host/whci/
H A Dasl.c29 struct whc_qset **next, struct whc_qset **prev) qset_get_next_prev()
38 p = qset->list_node.prev; qset_get_next_prev()
40 p = p->prev; qset_get_next_prev()
43 *prev = container_of(p, struct whc_qset, list_node); qset_get_next_prev()
55 struct whc_qset *next, *prev; asl_qset_insert() local
60 qset_get_next_prev(whc, qset, &next, &prev); asl_qset_insert()
62 whc_qset_set_link_ptr(&prev->qh.link, qset->qset_dma); asl_qset_insert()
68 struct whc_qset *prev, *next; asl_qset_remove() local
70 qset_get_next_prev(whc, qset, &next, &prev); asl_qset_remove()
83 whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma); asl_qset_remove()
28 qset_get_next_prev(struct whc *whc, struct whc_qset *qset, struct whc_qset **next, struct whc_qset **prev) qset_get_next_prev() argument
/linux-4.4.14/net/ipv4/
H A Dip_fragment.c98 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev,
336 struct sk_buff *prev, *next; ip_frag_queue() local
403 prev = qp->q.fragments_tail; ip_frag_queue()
404 if (!prev || FRAG_CB(prev)->offset < offset) { ip_frag_queue()
408 prev = NULL; ip_frag_queue()
412 prev = next; ip_frag_queue()
420 if (prev) { ip_frag_queue()
421 int i = (FRAG_CB(prev)->offset + prev->len) - offset; ip_frag_queue()
460 if (prev) ip_frag_queue()
461 prev->next = next; ip_frag_queue()
477 if (prev) ip_frag_queue()
478 prev->next = skb; ip_frag_queue()
508 err = ip_frag_reasm(qp, prev, dev); ip_frag_queue()
524 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, ip_frag_reasm() argument
543 if (prev) { ip_frag_reasm()
544 head = prev->next; ip_frag_reasm()
552 prev->next = fp; ip_frag_reasm()
/linux-4.4.14/drivers/staging/lustre/lustre/fld/
H A Dfld_cache.c226 curr = cache->fci_lru.prev; fld_cache_shrink()
232 curr = curr->prev; fld_cache_shrink()
308 * prev range only. fixup will handle overlapping with next range. */ fld_cache_overlap_handle()
343 fld_cache_entry_add(cache, f_new, f_curr->fce_list.prev); fld_cache_overlap_handle()
386 struct list_head *prev = NULL; fld_cache_insert_nolock() local
409 prev = &f_curr->fce_list; list_for_each_entry_safe()
418 if (prev == NULL)
419 prev = head;
423 fld_cache_entry_add(cache, f_new, prev);
520 struct fld_cache_entry *prev = NULL; fld_cache_lookup() local
529 if (prev != NULL) list_for_each_entry()
530 *range = prev->fce_range; list_for_each_entry()
534 prev = flde; list_for_each_entry()
/linux-4.4.14/drivers/iommu/
H A Diova.c105 struct rb_node *prev, *curr = NULL; __alloc_and_insert_iova_range() local
114 prev = curr; __alloc_and_insert_iova_range()
131 prev = curr; __alloc_and_insert_iova_range()
153 /* If we have 'prev', it's a valid place to start the __alloc_and_insert_iova_range()
155 if (prev) __alloc_and_insert_iova_range()
156 entry = &prev; __alloc_and_insert_iova_range()
517 struct iova *prev = NULL, *next = NULL; split_and_remove_iova() local
521 prev = alloc_and_init_iova(iova->pfn_lo, pfn_lo - 1); split_and_remove_iova()
522 if (prev == NULL) split_and_remove_iova()
534 if (prev) { split_and_remove_iova()
535 iova_insert_rbtree(&iovad->rbroot, prev); split_and_remove_iova()
548 if (prev) split_and_remove_iova()
549 free_iova_mem(prev); split_and_remove_iova()
/linux-4.4.14/arch/cris/arch-v10/kernel/
H A Dfasttimer.c260 t->prev = NULL; start_one_shot_timer()
264 fast_timer_list->prev = t; start_one_shot_timer()
279 t->prev = tmp; start_one_shot_timer()
283 tmp->next->prev = t; start_one_shot_timer()
296 return (t->next != NULL) || (t->prev != NULL) || (t == fast_timer_list); fast_timer_pending()
301 struct fast_timer *next, *prev; detach_fast_timer() local
305 prev = t->prev; detach_fast_timer()
307 next->prev = prev; detach_fast_timer()
308 if (prev) detach_fast_timer()
309 prev->next = next; detach_fast_timer()
323 t->next = t->prev = NULL; del_fast_timer()
384 if (t->prev) timer1_handler()
386 t->prev->next = t->next; timer1_handler()
394 t->next->prev = t->prev; timer1_handler()
396 t->prev = NULL; timer1_handler()
/linux-4.4.14/arch/cris/arch-v32/kernel/
H A Dfasttimer.c238 t->prev = NULL; start_one_shot_timer()
241 fast_timer_list->prev = t; start_one_shot_timer()
253 t->prev = tmp; start_one_shot_timer()
257 tmp->next->prev = t; start_one_shot_timer()
270 return (t->next != NULL) || (t->prev != NULL) || (t == fast_timer_list); fast_timer_pending()
275 struct fast_timer *next, *prev; detach_fast_timer() local
279 prev = t->prev; detach_fast_timer()
281 next->prev = prev; detach_fast_timer()
282 if (prev) detach_fast_timer()
283 prev->next = next; detach_fast_timer()
297 t->next = t->prev = NULL; del_fast_timer()
373 if (t->prev) timer_trig_handler()
374 t->prev->next = t->next; timer_trig_handler()
378 t->next->prev = t->prev; timer_trig_handler()
379 t->prev = NULL; timer_trig_handler()
/linux-4.4.14/drivers/mmc/core/
H A Dsdio_cis.c231 struct sdio_func_tuple *this, **prev; sdio_read_cis() local
255 prev = &func->tuples; sdio_read_cis()
257 prev = &card->tuples; sdio_read_cis()
259 BUG_ON(*prev); sdio_read_cis()
311 *prev = this; sdio_read_cis()
312 prev = &this->next; sdio_read_cis()
341 *prev = card->tuples; sdio_read_cis()
/linux-4.4.14/arch/sh/kernel/cpu/sh5/
H A Dswitchto.S24 r2 - prev
25 r3 - &prev->thread
30 r2 - last (=prev) : this just stays in r2 throughout
120 st.l r3, 0, r15 ! prev->thread.sp
121 st.l r3, 8, r1 ! prev->thread.kregs
122 st.l r3, 4, r9 ! prev->thread.pc
123 st.q r1, 0, r9 ! save prev->thread.pc into pt_regs->pc
/linux-4.4.14/arch/m68k/sun3/
H A Dsun3dvma.c110 struct hole *prev = NULL; refill() local
117 if(!prev) { refill()
118 prev = hole; refill()
122 if(hole->end == prev->start) { refill()
123 hole->size += prev->size; refill()
124 hole->end = prev->end; refill()
125 list_move(&(prev->list), &hole_cache); refill()
/linux-4.4.14/net/atm/
H A Dmpoa_caches.c112 entry->prev = NULL; in_cache_add_entry()
114 client->in_cache->prev = entry; in_cache_add_entry()
200 if (entry->prev != NULL) in_cache_remove_entry()
201 entry->prev->next = entry->next; in_cache_remove_entry()
205 entry->next->prev = entry->prev; in_cache_remove_entry()
435 if (entry->prev != NULL) eg_cache_remove_entry()
436 entry->prev->next = entry->next; eg_cache_remove_entry()
440 entry->next->prev = entry->prev; eg_cache_remove_entry()
475 entry->prev = NULL; eg_cache_add_entry()
477 client->eg_cache->prev = entry; eg_cache_add_entry()
H A Dmpoa_caches.h16 struct in_cache_entry *prev; member in struct:in_cache_entry
53 struct eg_cache_entry *prev; member in struct:eg_cache_entry
/linux-4.4.14/fs/autofs4/
H A Dexpire.c76 static struct dentry *get_next_positive_subdir(struct dentry *prev, get_next_positive_subdir() argument
86 if (prev) get_next_positive_subdir()
87 next = prev->d_child.next; get_next_positive_subdir()
89 prev = dget_dlock(root); get_next_positive_subdir()
90 next = prev->d_subdirs.next; get_next_positive_subdir()
97 dput(prev); get_next_positive_subdir()
115 dput(prev); get_next_positive_subdir()
123 static struct dentry *get_next_positive_dentry(struct dentry *prev, get_next_positive_dentry() argument
130 if (prev == NULL) get_next_positive_dentry()
135 p = prev; get_next_positive_dentry()
146 dput(prev); get_next_positive_dentry()
178 dput(prev); get_next_positive_dentry()
/linux-4.4.14/kernel/gcov/
H A Dbase.c137 struct gcov_info *prev = NULL; gcov_module_notifier() local
146 gcov_info_unlink(prev, info); gcov_module_notifier()
150 prev = info; gcov_module_notifier()
/linux-4.4.14/net/ipv6/
H A Dreassembly.c79 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev,
212 struct sk_buff *prev, *next; ip6_frag_queue() local
289 prev = fq->q.fragments_tail; ip6_frag_queue()
290 if (!prev || FRAG6_CB(prev)->offset < offset) { ip6_frag_queue()
294 prev = NULL; ip6_frag_queue()
298 prev = next; ip6_frag_queue()
310 if (prev && ip6_frag_queue()
311 (FRAG6_CB(prev)->offset + prev->len) > offset) ip6_frag_queue()
324 if (prev) ip6_frag_queue()
325 prev->next = skb; ip6_frag_queue()
353 res = ip6_frag_reasm(fq, prev, dev); ip6_frag_queue()
379 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, ip6_frag_reasm() argument
396 if (prev) { ip6_frag_reasm()
397 head = prev->next; ip6_frag_reasm()
406 prev->next = fp; ip6_frag_reasm()
/linux-4.4.14/drivers/md/bcache/
H A Dbset.c280 /* Space required for the prev pointers */ bset_prev_bytes()
293 kfree(t->prev); bch_btree_keys_free()
295 free_pages((unsigned long) t->prev, bch_btree_keys_free()
306 t->prev = NULL; bch_btree_keys_free()
330 t->prev = bset_prev_bytes(b) < PAGE_SIZE bch_btree_keys_alloc()
333 if (!t->prev) bch_btree_keys_alloc()
499 * bits we're going to store in bkey_float->mantissa. t->prev[j] stores the size
528 return (void *) (((uint64_t *) tree_to_bkey(t, j)) - t->prev[j]); tree_to_prev_bkey()
533 * maintain a full search tree, we just keep a simple lookup table in t->prev.
537 return cacheline_to_bkey(t, cacheline, t->prev[cacheline]); table_to_bkey()
596 t->prev = t[-1].prev + j; bset_alloc_tree()
613 t->prev[0] = bkey_to_cacheline_offset(t, 0, t->data->start); bch_bset_build_unwritten_tree()
637 struct bkey *prev = NULL, *k = t->data->start; bch_bset_build_written_tree() local
660 prev = k, k = bkey_next(k); bch_bset_build_written_tree()
662 t->prev[j] = bkey_u64s(prev); bch_bset_build_written_tree()
750 t->prev[j] += shift; bch_bset_fix_lookup_table()
752 if (t->prev[j] > 7) { bch_bset_fix_lookup_table()
758 t->prev[j] = bkey_to_cacheline_offset(t, j, k); bch_bset_fix_lookup_table()
771 t->prev[t->size] = bkey_to_cacheline_offset(t, t->size, k); bch_bset_fix_lookup_table()
824 struct bkey *m, *prev = NULL; bch_btree_insert_key() local
840 prev = m, m = bkey_next(m); bch_btree_insert_key()
842 /* prev is in the tree, if we merge we're done */ bch_btree_insert_key()
844 if (prev && bch_btree_insert_key()
845 bch_bkey_try_merge(b, prev, k)) bch_btree_insert_key()
/linux-4.4.14/drivers/scsi/
H A Dscsi_module.c47 while ((l = l->prev) != &sht->legacy_hosts) init_this_scsi_driver()
/linux-4.4.14/drivers/media/pci/cx88/
H A Dcx88-vbi.c168 struct cx88_buffer *prev; buffer_queue() local
184 prev = list_entry(q->active.prev, struct cx88_buffer, list); buffer_queue()
186 prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); buffer_queue()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/bios/
H A Dbase.c73 u8 *prev = bios->data; nvbios_extend() local
75 bios->data = prev; nvbios_extend()
78 memcpy(bios->data, prev, bios->size); nvbios_extend()
80 kfree(prev); nvbios_extend()
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/volt/
H A Dbase.c100 int prev = nvkm_volt_get(volt); nvkm_volt_set_id() local
101 if (!condition || prev < 0 || nvkm_volt_set_id()
102 (condition < 0 && ret < prev) || nvkm_volt_set_id()
103 (condition > 0 && ret > prev)) { nvkm_volt_set_id()
/linux-4.4.14/fs/
H A Duserfaultfd.c396 * list_del_init() to refile across the two lists, the prev handle_userfault()
430 struct vm_area_struct *vma, *prev; userfaultfd_release() local
446 prev = NULL; userfaultfd_release()
452 prev = vma; userfaultfd_release()
456 prev = vma_merge(mm, prev, vma->vm_start, vma->vm_end, userfaultfd_release()
461 if (prev) userfaultfd_release()
462 vma = prev; userfaultfd_release()
464 prev = vma; userfaultfd_release()
721 struct vm_area_struct *vma, *prev, *cur; userfaultfd_register() local
764 vma = find_vma_prev(mm, start, &prev); userfaultfd_register()
810 prev = vma; userfaultfd_register()
833 prev = vma_merge(mm, prev, start, vma_end, new_flags, userfaultfd_register()
837 if (prev) { userfaultfd_register()
838 vma = prev; userfaultfd_register()
861 prev = vma; userfaultfd_register()
885 struct vm_area_struct *vma, *prev, *cur; userfaultfd_unregister() local
906 vma = find_vma_prev(mm, start, &prev); userfaultfd_unregister()
947 prev = vma; userfaultfd_unregister()
967 prev = vma_merge(mm, prev, start, vma_end, new_flags, userfaultfd_unregister()
971 if (prev) { userfaultfd_unregister()
972 vma = prev; userfaultfd_unregister()
995 prev = vma; userfaultfd_unregister()
/linux-4.4.14/arch/sh/kernel/
H A Dprocess_32.c181 __switch_to(struct task_struct *prev, struct task_struct *next) __switch_to() argument
189 unlazy_fpu(prev, task_pt_regs(prev)); __switch_to()
213 return prev; __switch_to()
/linux-4.4.14/arch/arm64/include/asm/
H A Dmmu_context.h127 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
132 if (prev == next) switch_mm()
148 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
/linux-4.4.14/fs/hfsplus/
H A Dbnode.c308 be32_to_cpu(desc.next), be32_to_cpu(desc.prev), hfs_bnode_dump()
343 if (node->prev) { hfs_bnode_unlink()
344 tmp = hfs_bnode_find(tree, node->prev); hfs_bnode_unlink()
359 tmp->prev = node->prev; hfs_bnode_unlink()
360 cnid = cpu_to_be32(tmp->prev); hfs_bnode_unlink()
362 offsetof(struct hfs_bnode_desc, prev), 4); hfs_bnode_unlink()
365 tree->leaf_tail = node->prev; hfs_bnode_unlink()
368 if (!node->prev && !node->next) hfs_bnode_unlink()
510 node->prev = be32_to_cpu(desc->prev); hfs_bnode_find()
/linux-4.4.14/net/core/
H A Ddst.c322 unsigned long prev, new; dst_cow_metrics_generic() local
327 prev = cmpxchg(&dst->_metrics, old, new); dst_cow_metrics_generic()
329 if (prev != old) { dst_cow_metrics_generic()
331 p = __DST_METRICS_PTR(prev); dst_cow_metrics_generic()
332 if (prev & DST_METRICS_READ_ONLY) dst_cow_metrics_generic()
343 unsigned long prev, new; __dst_destroy_metrics_generic() local
346 prev = cmpxchg(&dst->_metrics, old, new); __dst_destroy_metrics_generic()
347 if (prev == old) __dst_destroy_metrics_generic()
/linux-4.4.14/tools/perf/util/
H A Dcpumap.c52 int n, cpu, prev; cpu_map__read() local
56 prev = -1; cpu_map__read()
61 if (prev >= 0) { cpu_map__read()
62 int new_max = nr_cpus + cpu - prev - 1; cpu_map__read()
72 while (++prev < cpu) cpu_map__read()
73 tmp_cpus[nr_cpus++] = prev; cpu_map__read()
85 prev = cpu; cpu_map__read()
87 prev = -1; cpu_map__read()
/linux-4.4.14/fs/nilfs2/
H A Dalloc.c212 struct nilfs_bh_assoc *prev, nilfs_palloc_get_block()
218 if (prev->bh && blkoff == prev->blkoff) { nilfs_palloc_get_block()
219 get_bh(prev->bh); nilfs_palloc_get_block()
220 *bhp = prev->bh; nilfs_palloc_get_block()
233 brelse(prev->bh); nilfs_palloc_get_block()
235 prev->bh = *bhp; nilfs_palloc_get_block()
236 prev->blkoff = blkoff; nilfs_palloc_get_block()
246 * @prev: nilfs_bh_assoc struct of the last used buffer
247 * @lock: spin lock protecting @prev
250 struct nilfs_bh_assoc *prev, nilfs_palloc_delete_block()
254 if (prev->bh && blkoff == prev->blkoff) { nilfs_palloc_delete_block()
255 brelse(prev->bh); nilfs_palloc_delete_block()
256 prev->bh = NULL; nilfs_palloc_delete_block()
206 nilfs_palloc_get_block(struct inode *inode, unsigned long blkoff, int create, void (*init_block)(struct inode *, struct buffer_head *, void *), struct buffer_head **bhp, struct nilfs_bh_assoc *prev, spinlock_t *lock) nilfs_palloc_get_block() argument
249 nilfs_palloc_delete_block(struct inode *inode, unsigned long blkoff, struct nilfs_bh_assoc *prev, spinlock_t *lock) nilfs_palloc_delete_block() argument
/linux-4.4.14/drivers/acpi/acpica/
H A Dpsloop.c319 union acpi_operand_object *prev; acpi_ps_link_module_code() local
328 prev = next = acpi_gbl_module_code_list; acpi_ps_link_module_code()
330 prev = next; acpi_ps_link_module_code()
338 if (!prev || acpi_ps_link_module_code()
339 ((prev->method.aml_start + prev->method.aml_length) != aml_start)) { acpi_ps_link_module_code()
370 if (!prev) { acpi_ps_link_module_code()
373 prev->method.mutex = method_obj; acpi_ps_link_module_code()
378 prev)); acpi_ps_link_module_code()
380 prev->method.aml_length += aml_length; acpi_ps_link_module_code()
/linux-4.4.14/arch/x86/kernel/
H A Dprocess_64.c277 struct thread_struct *prev = &prev_p->thread; __switch_to() local
279 struct fpu *prev_fpu = &prev->fpu; __switch_to()
325 savesegment(es, prev->es); __switch_to()
326 if (unlikely(next->es | prev->es)) __switch_to()
329 savesegment(ds, prev->ds); __switch_to()
330 if (unlikely(next->ds | prev->ds)) __switch_to()
368 if (unlikely(fsindex | next->fsindex | prev->fs)) { __switch_to()
381 prev->fs = 0; __switch_to()
385 prev->fsindex = fsindex; __switch_to()
387 if (unlikely(gsindex | next->gsindex | prev->gs)) { __switch_to()
392 prev->gs = 0; __switch_to()
396 prev->gsindex = gsindex; __switch_to()
422 prev->iopl != next->iopl)) __switch_to()

Completed in 4930 milliseconds

123456