/linux-4.4.14/lib/ |
H A D | list_debug.c | 19 * the prev/next entries already! 23 struct list_head *prev, __list_add() 26 WARN(next->prev != prev, __list_add() 27 "list_add corruption. next->prev should be " __list_add() 28 "prev (%p), but was %p. (next=%p).\n", __list_add() 29 prev, next->prev, next); __list_add() 30 WARN(prev->next != next, __list_add() 31 "list_add corruption. prev->next should be " __list_add() 32 "next (%p), but was %p. (prev=%p).\n", __list_add() 33 next, prev->next, prev); __list_add() 34 WARN(new == prev || new == next, __list_add() 35 "list_add double add: new=%p, prev=%p, next=%p.\n", __list_add() 36 new, prev, next); __list_add() 37 next->prev = new; __list_add() 39 new->prev = prev; __list_add() 40 prev->next = new; __list_add() 46 struct list_head *prev, *next; __list_del_entry() local 48 prev = entry->prev; __list_del_entry() 54 WARN(prev == LIST_POISON2, __list_del_entry() 55 "list_del corruption, %p->prev is LIST_POISON2 (%p)\n", __list_del_entry() 57 WARN(prev->next != entry, __list_del_entry() 58 "list_del corruption. prev->next should be %p, " __list_del_entry() 59 "but was %p\n", entry, prev->next) || __list_del_entry() 60 WARN(next->prev != entry, __list_del_entry() 61 "list_del corruption. next->prev should be %p, " __list_del_entry() 62 "but was %p\n", entry, next->prev)) __list_del_entry() 65 __list_del(prev, next); __list_del_entry() 79 entry->prev = LIST_POISON2; list_del() 87 struct list_head *prev, struct list_head *next) __list_add_rcu() 89 WARN(next->prev != prev, __list_add_rcu() 90 "list_add_rcu corruption. next->prev should be prev (%p), but was %p. (next=%p).\n", __list_add_rcu() 91 prev, next->prev, next); __list_add_rcu() 92 WARN(prev->next != next, __list_add_rcu() 93 "list_add_rcu corruption. prev->next should be next (%p), but was %p. (prev=%p).\n", __list_add_rcu() 94 next, prev->next, prev); __list_add_rcu() 96 new->prev = prev; __list_add_rcu() 97 rcu_assign_pointer(list_next_rcu(prev), new); __list_add_rcu() 98 next->prev = new; __list_add_rcu() 22 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 86 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
|
H A D | plist.c | 36 WARN(n->prev != p || p->next != n, plist_check_prev_next() 38 "prev: %p, n: %p, p: %p\n" plist_check_prev_next() 40 t, t->next, t->prev, plist_check_prev_next() 41 p, p->next, p->prev, plist_check_prev_next() 42 n, n->next, n->prev); plist_check_prev_next() 47 struct list_head *prev = top, *next = top->next; plist_check_list() local 49 plist_check_prev_next(top, prev, next); plist_check_list() 51 prev = next; plist_check_list() 52 next = prev->next; plist_check_list() 53 plist_check_prev_next(top, prev, next); plist_check_list() 76 struct plist_node *first, *iter, *prev = NULL; plist_add() local 94 prev = iter; plist_add() 99 if (!prev || prev->prio != node->prio) plist_add()
|
H A D | klist.c | 327 * klist_prev - Ante up prev node in list. 331 * node, if there was one. Grab the prev node, increment its reference 332 * count, drop the lock, and return that prev node. 338 struct klist_node *prev; klist_prev() local 343 prev = to_klist_node(last->n_node.prev); klist_prev() 347 prev = to_klist_node(i->i_klist->k_list.prev); klist_prev() 350 while (prev != to_klist_node(&i->i_klist->k_list)) { klist_prev() 351 if (likely(!knode_dead(prev))) { klist_prev() 352 kref_get(&prev->n_ref); klist_prev() 353 i->i_cur = prev; klist_prev() 356 prev = to_klist_node(prev->n_node.prev); klist_prev()
|
H A D | list_sort.c | 17 * sentinel head node, "prev" links not maintained. 45 * prev-link restoration pass, or maintaining the prev links 61 a->prev = tail; merge_and_restore_back_links() 65 b->prev = tail; merge_and_restore_back_links() 82 tail->next->prev = tail; merge_and_restore_back_links() 87 head->prev = tail; merge_and_restore_back_links() 119 head->prev->next = NULL; list_sort() 249 if (cur->next->prev != cur) { list_sort_test() 274 if (head.prev != cur) { list_sort_test()
|
/linux-4.4.14/arch/score/include/asm/ |
H A D | switch_to.h | 6 #define switch_to(prev, next, last) \ 8 (last) = resume(prev, next, task_thread_info(next)); \
|
H A D | mmu_context.h | 72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 102 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
|
/linux-4.4.14/arch/parisc/include/asm/ |
H A D | switch_to.h | 8 #define switch_to(prev, next, last) do { \ 9 (last) = _switch_to(prev, next); \
|
H A D | mmu_context.h | 52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument 55 if (prev != next) { switch_mm() 63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 78 switch_mm(prev,next,current); activate_mm()
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | switch_to.h | 31 #define switch_to(prev,next,last) do { \ 32 if (prev->mm) { \ 34 save_access_regs(&prev->thread.acrs[0]); \ 35 save_ri_cb(prev->thread.ri_cb); \ 41 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \ 43 prev = __switch_to(prev,next); \
|
H A D | mmu_context.h | 85 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 91 if (prev == next) switch_mm() 99 atomic_dec(&prev->context.attach_count); switch_mm() 101 cpumask_clear_cpu(cpu, &prev->context.cpu_attach_mask); switch_mm() 127 static inline void activate_mm(struct mm_struct *prev, activate_mm() argument 130 switch_mm(prev, next, current); activate_mm()
|
/linux-4.4.14/tools/firewire/ |
H A D | list.h | 2 struct list *next, *prev; member in struct:list 9 list->prev = list; list_init() 21 new_link->prev = link->prev; list_insert() 23 new_link->prev->next = new_link; list_insert() 24 new_link->next->prev = new_link; list_insert() 42 link->prev->next = link->next; list_remove() 43 link->next->prev = link->prev; list_remove() 53 list_entry((list)->prev, type, member)
|
H A D | nosy-dump.c | 498 struct subaction *sa, *prev; handle_request_packet() local 507 prev = list_tail(&t->request_list, handle_request_packet() 510 if (!ACK_BUSY(prev->ack)) { handle_request_packet() 519 if (prev->packet.common.tcode != sa->packet.common.tcode || handle_request_packet() 520 prev->packet.common.tlabel != sa->packet.common.tlabel) { handle_request_packet() 563 struct subaction *sa, *prev; handle_response_packet() local 576 prev = list_tail(&t->response_list, struct subaction, link); handle_response_packet() 578 if (!ACK_BUSY(prev->ack)) { handle_response_packet() 585 if (prev->packet.common.tcode != sa->packet.common.tcode || handle_response_packet() 586 prev->packet.common.tlabel != sa->packet.common.tlabel) { handle_response_packet() 591 prev = list_tail(&t->request_list, struct subaction, link); handle_response_packet() 592 if (prev->ack != ACK_PENDING) { handle_response_packet() 599 if (packet_info[prev->packet.common.tcode].response_tcode != handle_response_packet()
|
/linux-4.4.14/include/asm-generic/ |
H A D | cmpxchg-local.h | 17 unsigned long flags, prev; __cmpxchg_local_generic() local 27 case 1: prev = *(u8 *)ptr; __cmpxchg_local_generic() 28 if (prev == old) __cmpxchg_local_generic() 31 case 2: prev = *(u16 *)ptr; __cmpxchg_local_generic() 32 if (prev == old) __cmpxchg_local_generic() 35 case 4: prev = *(u32 *)ptr; __cmpxchg_local_generic() 36 if (prev == old) __cmpxchg_local_generic() 39 case 8: prev = *(u64 *)ptr; __cmpxchg_local_generic() 40 if (prev == old) __cmpxchg_local_generic() 47 return prev; __cmpxchg_local_generic() 56 u64 prev; __cmpxchg64_local_generic() local 60 prev = *(u64 *)ptr; __cmpxchg64_local_generic() 61 if (prev == old) __cmpxchg64_local_generic() 64 return prev; __cmpxchg64_local_generic()
|
H A D | mutex-xchg.h | 94 int prev = atomic_xchg_acquire(count, 0); __mutex_fastpath_trylock() local 96 if (unlikely(prev < 0)) { __mutex_fastpath_trylock() 99 * state. If while doing so we get back a prev value of 1 __mutex_fastpath_trylock() 108 prev = atomic_xchg_acquire(count, prev); __mutex_fastpath_trylock() 109 if (prev < 0) __mutex_fastpath_trylock() 110 prev = 0; __mutex_fastpath_trylock() 113 return prev; __mutex_fastpath_trylock()
|
H A D | switch_to.h | 25 #define switch_to(prev, next, last) \ 27 ((last) = __switch_to((prev), (next))); \
|
H A D | mmu_context.h | 34 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument
|
/linux-4.4.14/arch/unicore32/include/asm/ |
H A D | switch_to.h | 17 * switch_to(prev, next) should switch from task `prev' to `next' 18 * `prev' will never be the same as `next'. schedule() itself 24 #define switch_to(prev, next, last) \ 26 last = __switch_to(prev, task_thread_info(prev), \
|
H A D | mmu_context.h | 49 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 54 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) switch_mm() 59 #define activate_mm(prev, next) switch_mm(prev, next, NULL)
|
/linux-4.4.14/arch/blackfin/include/asm/ |
H A D | switch_to.h | 21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next); 24 #define switch_to(prev,next,last) \ 26 memcpy (&task_thread_info(prev)->l1_task_info, L1_SCRATCH_TASK_INFO, \ 30 (last) = resume (prev, next); \ 33 #define switch_to(prev, next, last) \ 35 (last) = resume(prev, next); \
|
H A D | mmu_context.h | 71 #define activate_mm(prev, next) switch_mm(prev, next, NULL) 111 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 116 __switch_mm(prev, next, tsk); switch_mm() 160 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 163 __switch_mm(prev, next, tsk); switch_mm()
|
/linux-4.4.14/drivers/media/platform/omap3isp/ |
H A D | isppreview.c | 152 preview_config_luma_enhancement(struct isp_prev_device *prev, preview_config_luma_enhancement() argument 155 struct isp_device *isp = to_isp_device(prev); preview_config_luma_enhancement() 171 preview_enable_luma_enhancement(struct isp_prev_device *prev, bool enable) preview_enable_luma_enhancement() argument 173 struct isp_device *isp = to_isp_device(prev); preview_enable_luma_enhancement() 186 static void preview_enable_invalaw(struct isp_prev_device *prev, bool enable) preview_enable_invalaw() argument 188 struct isp_device *isp = to_isp_device(prev); preview_enable_invalaw() 201 static void preview_config_hmed(struct isp_prev_device *prev, preview_config_hmed() argument 204 struct isp_device *isp = to_isp_device(prev); preview_config_hmed() 216 static void preview_enable_hmed(struct isp_prev_device *prev, bool enable) preview_enable_hmed() argument 218 struct isp_device *isp = to_isp_device(prev); preview_enable_hmed() 236 static void preview_config_cfa(struct isp_prev_device *prev, preview_config_cfa() argument 245 const unsigned int *order = cfa_coef_order[prev->params.cfa_order]; preview_config_cfa() 247 struct isp_device *isp = to_isp_device(prev); preview_config_cfa() 272 preview_config_chroma_suppression(struct isp_prev_device *prev, preview_config_chroma_suppression() argument 275 struct isp_device *isp = to_isp_device(prev); preview_config_chroma_suppression() 288 preview_enable_chroma_suppression(struct isp_prev_device *prev, bool enable) preview_enable_chroma_suppression() argument 290 struct isp_device *isp = to_isp_device(prev); preview_enable_chroma_suppression() 306 preview_config_whitebalance(struct isp_prev_device *prev, preview_config_whitebalance() argument 309 struct isp_device *isp = to_isp_device(prev); preview_config_whitebalance() 345 preview_config_blkadj(struct isp_prev_device *prev, preview_config_blkadj() argument 348 struct isp_device *isp = to_isp_device(prev); preview_config_blkadj() 361 preview_config_rgb_blending(struct isp_prev_device *prev, preview_config_rgb_blending() argument 364 struct isp_device *isp = to_isp_device(prev); preview_config_rgb_blending() 399 preview_config_csc(struct isp_prev_device *prev, preview_config_csc() argument 402 struct isp_device *isp = to_isp_device(prev); preview_config_csc() 431 preview_config_yc_range(struct isp_prev_device *prev, preview_config_yc_range() argument 434 struct isp_device *isp = to_isp_device(prev); preview_config_yc_range() 449 preview_config_dcor(struct isp_prev_device *prev, preview_config_dcor() argument 452 struct isp_device *isp = to_isp_device(prev); preview_config_dcor() 471 static void preview_enable_dcor(struct isp_prev_device *prev, bool enable) preview_enable_dcor() argument 473 struct isp_device *isp = to_isp_device(prev); preview_enable_dcor() 487 preview_enable_drkframe_capture(struct isp_prev_device *prev, bool enable) preview_enable_drkframe_capture() argument 489 struct isp_device *isp = to_isp_device(prev); preview_enable_drkframe_capture() 502 static void preview_enable_drkframe(struct isp_prev_device *prev, bool enable) preview_enable_drkframe() argument 504 struct isp_device *isp = to_isp_device(prev); preview_enable_drkframe() 518 preview_config_noisefilter(struct isp_prev_device *prev, preview_config_noisefilter() argument 521 struct isp_device *isp = to_isp_device(prev); preview_config_noisefilter() 538 preview_enable_noisefilter(struct isp_prev_device *prev, bool enable) preview_enable_noisefilter() argument 540 struct isp_device *isp = to_isp_device(prev); preview_enable_noisefilter() 554 preview_config_gammacorrn(struct isp_prev_device *prev, preview_config_gammacorrn() argument 557 struct isp_device *isp = to_isp_device(prev); preview_config_gammacorrn() 587 preview_enable_gammacorrn(struct isp_prev_device *prev, bool enable) preview_enable_gammacorrn() argument 589 struct isp_device *isp = to_isp_device(prev); preview_enable_gammacorrn() 605 preview_config_contrast(struct isp_prev_device *prev, preview_config_contrast() argument 608 struct isp_device *isp = to_isp_device(prev); preview_config_contrast() 619 preview_config_brightness(struct isp_prev_device *prev, preview_config_brightness() argument 622 struct isp_device *isp = to_isp_device(prev); preview_config_brightness() 636 preview_update_contrast(struct isp_prev_device *prev, u8 contrast) preview_update_contrast() argument 641 spin_lock_irqsave(&prev->params.lock, flags); preview_update_contrast() 642 params = (prev->params.active & OMAP3ISP_PREV_CONTRAST) preview_update_contrast() 643 ? &prev->params.params[0] : &prev->params.params[1]; preview_update_contrast() 649 spin_unlock_irqrestore(&prev->params.lock, flags); preview_update_contrast() 658 preview_update_brightness(struct isp_prev_device *prev, u8 brightness) preview_update_brightness() argument 663 spin_lock_irqsave(&prev->params.lock, flags); preview_update_brightness() 664 params = (prev->params.active & OMAP3ISP_PREV_BRIGHTNESS) preview_update_brightness() 665 ? &prev->params.params[0] : &prev->params.params[1]; preview_update_brightness() 671 spin_unlock_irqrestore(&prev->params.lock, flags); preview_update_brightness() 675 preview_params_lock(struct isp_prev_device *prev, u32 update, bool shadow) preview_params_lock() argument 677 u32 active = prev->params.active; preview_params_lock() 681 prev->params.params[0].busy |= ~active & update; preview_params_lock() 682 prev->params.params[1].busy |= active & update; preview_params_lock() 685 update = (prev->params.params[0].update & active) preview_params_lock() 686 | (prev->params.params[1].update & ~active); preview_params_lock() 688 prev->params.params[0].busy |= active & update; preview_params_lock() 689 prev->params.params[1].busy |= ~active & update; preview_params_lock() 696 preview_params_unlock(struct isp_prev_device *prev, u32 update, bool shadow) preview_params_unlock() argument 698 u32 active = prev->params.active; preview_params_unlock() 704 prev->params.params[0].update |= (~active & update); preview_params_unlock() 705 prev->params.params[1].update |= (active & update); preview_params_unlock() 706 prev->params.params[0].busy &= active; preview_params_unlock() 707 prev->params.params[1].busy &= ~active; preview_params_unlock() 712 prev->params.params[0].update &= ~(active & update); preview_params_unlock() 713 prev->params.params[1].update &= ~(~active & update); preview_params_unlock() 714 prev->params.params[0].busy &= ~active; preview_params_unlock() 715 prev->params.params[1].busy &= active; preview_params_unlock() 719 static void preview_params_switch(struct isp_prev_device *prev) preview_params_switch() argument 727 to_switch = (prev->params.params[0].update & ~prev->params.active) preview_params_switch() 728 | (prev->params.params[1].update & prev->params.active); preview_params_switch() 729 to_switch &= ~(prev->params.params[0].busy | preview_params_switch() 730 prev->params.params[1].busy); preview_params_switch() 734 prev->params.active ^= to_switch; preview_params_switch() 739 prev->params.params[0].update &= ~(~prev->params.active & to_switch); preview_params_switch() 740 prev->params.params[1].update &= ~(prev->params.active & to_switch); preview_params_switch() 856 * @prev: ISP preview engine 862 static int preview_config(struct isp_prev_device *prev, preview_config() argument 875 spin_lock_irqsave(&prev->params.lock, flags); preview_config() 876 preview_params_lock(prev, cfg->update, true); preview_config() 877 active = prev->params.active; preview_config() 878 spin_unlock_irqrestore(&prev->params.lock, flags); preview_config() 890 params = &prev->params.params[!!(active & bit)]; preview_config() 912 spin_lock_irqsave(&prev->params.lock, flags); preview_config() 913 preview_params_unlock(prev, update, true); preview_config() 914 preview_params_switch(prev); preview_config() 915 spin_unlock_irqrestore(&prev->params.lock, flags); preview_config() 922 * @prev: pointer to preview private structure 928 static void preview_setup_hw(struct isp_prev_device *prev, u32 update, preview_setup_hw() argument 944 params = &prev->params.params[!(active & bit)]; preview_setup_hw() 948 attr->config(prev, params); preview_setup_hw() 950 attr->enable(prev, true); preview_setup_hw() 953 attr->enable(prev, false); preview_setup_hw() 960 * @prev: pointer to previewer private structure 963 static void preview_config_ycpos(struct isp_prev_device *prev, u32 pixelcode) preview_config_ycpos() argument 965 struct isp_device *isp = to_isp_device(prev); preview_config_ycpos() 988 static void preview_config_averager(struct isp_prev_device *prev, u8 average) preview_config_averager() argument 990 struct isp_device *isp = to_isp_device(prev); preview_config_averager() 1000 * @prev: The preview engine 1011 static void preview_config_input_format(struct isp_prev_device *prev, preview_config_input_format() argument 1014 struct isp_device *isp = to_isp_device(prev); preview_config_input_format() 1026 prev->params.cfa_order = 0; preview_config_input_format() 1029 prev->params.cfa_order = 1; preview_config_input_format() 1032 prev->params.cfa_order = 2; preview_config_input_format() 1035 prev->params.cfa_order = 3; preview_config_input_format() 1048 params = (prev->params.active & OMAP3ISP_PREV_CFA) preview_config_input_format() 1049 ? &prev->params.params[0] : &prev->params.params[1]; preview_config_input_format() 1051 preview_config_cfa(prev, params); preview_config_input_format() 1065 static void preview_config_input_size(struct isp_prev_device *prev, u32 active) preview_config_input_size() argument 1067 const struct v4l2_mbus_framefmt *format = &prev->formats[PREV_PAD_SINK]; preview_config_input_size() 1068 struct isp_device *isp = to_isp_device(prev); preview_config_input_size() 1069 unsigned int sph = prev->crop.left; preview_config_input_size() 1070 unsigned int eph = prev->crop.left + prev->crop.width - 1; preview_config_input_size() 1071 unsigned int slv = prev->crop.top; preview_config_input_size() 1072 unsigned int elv = prev->crop.top + prev->crop.height - 1; preview_config_input_size() 1083 features = (prev->params.params[0].features & active) preview_config_input_size() 1084 | (prev->params.params[1].features & ~active); preview_config_input_size() 1107 * @prev: Preview module 1116 preview_config_inlineoffset(struct isp_prev_device *prev, u32 offset) preview_config_inlineoffset() argument 1118 struct isp_device *isp = to_isp_device(prev); preview_config_inlineoffset() 1130 static void preview_set_inaddr(struct isp_prev_device *prev, u32 addr) preview_set_inaddr() argument 1132 struct isp_device *isp = to_isp_device(prev); preview_set_inaddr() 1143 static void preview_config_outlineoffset(struct isp_prev_device *prev, preview_config_outlineoffset() argument 1146 struct isp_device *isp = to_isp_device(prev); preview_config_outlineoffset() 1158 static void preview_set_outaddr(struct isp_prev_device *prev, u32 addr) preview_set_outaddr() argument 1160 struct isp_device *isp = to_isp_device(prev); preview_set_outaddr() 1165 static void preview_adjust_bandwidth(struct isp_prev_device *prev) preview_adjust_bandwidth() argument 1167 struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); preview_adjust_bandwidth() 1168 struct isp_device *isp = to_isp_device(prev); preview_adjust_bandwidth() 1169 const struct v4l2_mbus_framefmt *ifmt = &prev->formats[PREV_PAD_SINK]; preview_adjust_bandwidth() 1179 if (prev->input != PREVIEW_INPUT_MEMORY) { preview_adjust_bandwidth() 1218 int omap3isp_preview_busy(struct isp_prev_device *prev) omap3isp_preview_busy() argument 1220 struct isp_device *isp = to_isp_device(prev); omap3isp_preview_busy() 1231 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_restore_context() local 1234 prev->params.params[0].update = prev->params.active & update; omap3isp_preview_restore_context() 1235 prev->params.params[1].update = ~prev->params.active & update; omap3isp_preview_restore_context() 1237 preview_setup_hw(prev, update, prev->params.active); omap3isp_preview_restore_context() 1239 prev->params.params[0].update = 0; omap3isp_preview_restore_context() 1240 prev->params.params[1].update = 0; omap3isp_preview_restore_context() 1250 static void preview_print_status(struct isp_prev_device *prev) preview_print_status() argument 1252 struct isp_device *isp = to_isp_device(prev); preview_print_status() 1298 * @prev: pointer to previewer private structure 1300 static void preview_init_params(struct isp_prev_device *prev) preview_init_params() argument 1305 spin_lock_init(&prev->params.lock); preview_init_params() 1307 prev->params.active = ~0; preview_init_params() 1308 prev->params.params[0].busy = 0; preview_init_params() 1309 prev->params.params[0].update = OMAP3ISP_PREV_FEATURES_END - 1; preview_init_params() 1310 prev->params.params[1].busy = 0; preview_init_params() 1311 prev->params.params[1].update = 0; preview_init_params() 1313 params = &prev->params.params[0]; preview_init_params() 1361 * @prev: pointer to previewer private structure 1364 static unsigned int preview_max_out_width(struct isp_prev_device *prev) preview_max_out_width() argument 1366 struct isp_device *isp = to_isp_device(prev); preview_max_out_width() 1381 static void preview_configure(struct isp_prev_device *prev) preview_configure() argument 1383 struct isp_device *isp = to_isp_device(prev); preview_configure() 1390 spin_lock_irqsave(&prev->params.lock, flags); preview_configure() 1392 update = preview_params_lock(prev, 0, false); preview_configure() 1393 active = prev->params.active; preview_configure() 1394 spin_unlock_irqrestore(&prev->params.lock, flags); preview_configure() 1397 format = &prev->formats[PREV_PAD_SINK]; preview_configure() 1400 preview_adjust_bandwidth(prev); preview_configure() 1402 preview_config_input_format(prev, info); preview_configure() 1403 preview_config_input_size(prev, active); preview_configure() 1405 if (prev->input == PREVIEW_INPUT_CCDC) preview_configure() 1406 preview_config_inlineoffset(prev, 0); preview_configure() 1408 preview_config_inlineoffset(prev, ALIGN(format->width, 0x20) * preview_configure() 1411 preview_setup_hw(prev, update, active); preview_configure() 1414 format = &prev->formats[PREV_PAD_SOURCE]; preview_configure() 1416 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_configure() 1423 if (prev->output & PREVIEW_OUTPUT_RESIZER) preview_configure() 1430 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_configure() 1431 preview_config_outlineoffset(prev, preview_configure() 1434 preview_config_averager(prev, 0); preview_configure() 1435 preview_config_ycpos(prev, format->code); preview_configure() 1437 spin_lock_irqsave(&prev->params.lock, flags); preview_configure() 1438 preview_params_unlock(prev, update, false); preview_configure() 1439 spin_unlock_irqrestore(&prev->params.lock, flags); preview_configure() 1446 static void preview_enable_oneshot(struct isp_prev_device *prev) preview_enable_oneshot() argument 1448 struct isp_device *isp = to_isp_device(prev); preview_enable_oneshot() 1454 if (prev->input == PREVIEW_INPUT_MEMORY) preview_enable_oneshot() 1462 void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev) omap3isp_preview_isr_frame_sync() argument 1470 if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS && omap3isp_preview_isr_frame_sync() 1471 prev->video_out.dmaqueue_flags & ISP_VIDEO_DMAQUEUE_QUEUED) { omap3isp_preview_isr_frame_sync() 1472 preview_enable_oneshot(prev); omap3isp_preview_isr_frame_sync() 1473 isp_video_dmaqueue_flags_clr(&prev->video_out); omap3isp_preview_isr_frame_sync() 1477 static void preview_isr_buffer(struct isp_prev_device *prev) preview_isr_buffer() argument 1479 struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); preview_isr_buffer() 1483 if (prev->input == PREVIEW_INPUT_MEMORY) { preview_isr_buffer() 1484 buffer = omap3isp_video_buffer_next(&prev->video_in); preview_isr_buffer() 1486 preview_set_inaddr(prev, buffer->dma); preview_isr_buffer() 1490 if (prev->output & PREVIEW_OUTPUT_MEMORY) { preview_isr_buffer() 1491 buffer = omap3isp_video_buffer_next(&prev->video_out); preview_isr_buffer() 1493 preview_set_outaddr(prev, buffer->dma); preview_isr_buffer() 1499 switch (prev->state) { preview_isr_buffer() 1511 preview_enable_oneshot(prev); preview_isr_buffer() 1525 void omap3isp_preview_isr(struct isp_prev_device *prev) omap3isp_preview_isr() argument 1531 if (omap3isp_module_sync_is_stopping(&prev->wait, &prev->stopping)) omap3isp_preview_isr() 1534 spin_lock_irqsave(&prev->params.lock, flags); omap3isp_preview_isr() 1535 preview_params_switch(prev); omap3isp_preview_isr() 1536 update = preview_params_lock(prev, 0, false); omap3isp_preview_isr() 1537 active = prev->params.active; omap3isp_preview_isr() 1538 spin_unlock_irqrestore(&prev->params.lock, flags); omap3isp_preview_isr() 1540 preview_setup_hw(prev, update, active); omap3isp_preview_isr() 1541 preview_config_input_size(prev, active); omap3isp_preview_isr() 1543 if (prev->input == PREVIEW_INPUT_MEMORY || omap3isp_preview_isr() 1544 prev->output & PREVIEW_OUTPUT_MEMORY) omap3isp_preview_isr() 1545 preview_isr_buffer(prev); omap3isp_preview_isr() 1546 else if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS) omap3isp_preview_isr() 1547 preview_enable_oneshot(prev); omap3isp_preview_isr() 1549 spin_lock_irqsave(&prev->params.lock, flags); omap3isp_preview_isr() 1550 preview_params_unlock(prev, update, false); omap3isp_preview_isr() 1551 spin_unlock_irqrestore(&prev->params.lock, flags); omap3isp_preview_isr() 1561 struct isp_prev_device *prev = &video->isp->isp_prev; preview_video_queue() local 1564 preview_set_inaddr(prev, buffer->dma); preview_video_queue() 1567 preview_set_outaddr(prev, buffer->dma); preview_video_queue() 1586 struct isp_prev_device *prev = preview_s_ctrl() local 1591 preview_update_brightness(prev, ctrl->val); preview_s_ctrl() 1594 preview_update_contrast(prev, ctrl->val); preview_s_ctrl() 1614 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_ioctl() local 1618 return preview_config(prev, arg); preview_ioctl() 1633 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_stream() local 1634 struct isp_video *video_out = &prev->video_out; preview_set_stream() 1635 struct isp_device *isp = to_isp_device(prev); preview_set_stream() 1636 struct device *dev = to_device(prev); preview_set_stream() 1638 if (prev->state == ISP_PIPELINE_STREAM_STOPPED) { preview_set_stream() 1643 preview_configure(prev); preview_set_stream() 1644 atomic_set(&prev->stopping, 0); preview_set_stream() 1645 preview_print_status(prev); preview_set_stream() 1650 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_set_stream() 1654 !(prev->output & PREVIEW_OUTPUT_MEMORY)) preview_set_stream() 1655 preview_enable_oneshot(prev); preview_set_stream() 1661 if (prev->input == PREVIEW_INPUT_MEMORY) preview_set_stream() 1663 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_set_stream() 1666 preview_enable_oneshot(prev); preview_set_stream() 1670 if (omap3isp_module_sync_idle(&sd->entity, &prev->wait, preview_set_stream() 1671 &prev->stopping)) preview_set_stream() 1680 prev->state = enable; preview_set_stream() 1685 __preview_get_format(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, __preview_get_format() argument 1689 return v4l2_subdev_get_try_format(&prev->subdev, cfg, pad); __preview_get_format() 1691 return &prev->formats[pad]; __preview_get_format() 1695 __preview_get_crop(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, __preview_get_crop() argument 1699 return v4l2_subdev_get_try_crop(&prev->subdev, cfg, PREV_PAD_SINK); __preview_get_crop() 1701 return &prev->crop; __preview_get_crop() 1725 * @prev: ISP preview engine 1734 static void preview_try_format(struct isp_prev_device *prev, preview_try_format() argument 1754 if (prev->input == PREVIEW_INPUT_MEMORY) { preview_try_format() 1756 preview_max_out_width(prev)); preview_try_format() 1776 *fmt = *__preview_get_format(prev, cfg, PREV_PAD_SINK, which); preview_try_format() 1794 crop = __preview_get_crop(prev, cfg, which); preview_try_format() 1807 * @prev: ISP preview engine 1817 static void preview_try_crop(struct isp_prev_device *prev, preview_try_crop() argument 1831 if (prev->input == PREVIEW_INPUT_CCDC) { preview_try_crop() 1895 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_enum_frame_size() local 1904 preview_try_format(prev, cfg, fse->pad, &format, fse->which); preview_enum_frame_size() 1914 preview_try_format(prev, cfg, fse->pad, &format, fse->which); preview_enum_frame_size() 1935 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_get_selection() local 1948 format = __preview_get_format(prev, cfg, PREV_PAD_SINK, preview_get_selection() 1950 preview_try_crop(prev, format, &sel->r); preview_get_selection() 1954 sel->r = *__preview_get_crop(prev, cfg, sel->which); preview_get_selection() 1978 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_selection() local 1986 if (prev->state != ISP_PIPELINE_STREAM_STOPPED) preview_set_selection() 1994 sel->r = *__preview_get_crop(prev, cfg, sel->which); preview_set_selection() 1998 format = __preview_get_format(prev, cfg, PREV_PAD_SINK, sel->which); preview_set_selection() 1999 preview_try_crop(prev, format, &sel->r); preview_set_selection() 2000 *__preview_get_crop(prev, cfg, sel->which) = sel->r; preview_set_selection() 2003 format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, sel->which); preview_set_selection() 2004 preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, sel->which); preview_set_selection() 2019 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_get_format() local 2022 format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); preview_get_format() 2040 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_format() local 2044 format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); preview_set_format() 2048 preview_try_format(prev, cfg, fmt->pad, &fmt->format, fmt->which); preview_set_format() 2054 crop = __preview_get_crop(prev, cfg, fmt->which); preview_set_format() 2060 preview_try_crop(prev, &fmt->format, crop); preview_set_format() 2063 format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, preview_set_format() 2065 preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, preview_set_format() 2146 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_link_setup() local 2152 if (prev->input == PREVIEW_INPUT_CCDC) preview_link_setup() 2154 prev->input = PREVIEW_INPUT_MEMORY; preview_link_setup() 2156 if (prev->input == PREVIEW_INPUT_MEMORY) preview_link_setup() 2157 prev->input = PREVIEW_INPUT_NONE; preview_link_setup() 2164 if (prev->input == PREVIEW_INPUT_MEMORY) preview_link_setup() 2166 prev->input = PREVIEW_INPUT_CCDC; preview_link_setup() 2168 if (prev->input == PREVIEW_INPUT_CCDC) preview_link_setup() 2169 prev->input = PREVIEW_INPUT_NONE; preview_link_setup() 2181 if (prev->output & ~PREVIEW_OUTPUT_MEMORY) preview_link_setup() 2183 prev->output |= PREVIEW_OUTPUT_MEMORY; preview_link_setup() 2185 prev->output &= ~PREVIEW_OUTPUT_MEMORY; preview_link_setup() 2192 if (prev->output & ~PREVIEW_OUTPUT_RESIZER) preview_link_setup() 2194 prev->output |= PREVIEW_OUTPUT_RESIZER; preview_link_setup() 2196 prev->output &= ~PREVIEW_OUTPUT_RESIZER; preview_link_setup() 2213 void omap3isp_preview_unregister_entities(struct isp_prev_device *prev) omap3isp_preview_unregister_entities() argument 2215 v4l2_device_unregister_subdev(&prev->subdev); omap3isp_preview_unregister_entities() 2216 omap3isp_video_unregister(&prev->video_in); omap3isp_preview_unregister_entities() 2217 omap3isp_video_unregister(&prev->video_out); omap3isp_preview_unregister_entities() 2220 int omap3isp_preview_register_entities(struct isp_prev_device *prev, omap3isp_preview_register_entities() argument 2226 ret = v4l2_device_register_subdev(vdev, &prev->subdev); omap3isp_preview_register_entities() 2230 ret = omap3isp_video_register(&prev->video_in, vdev); omap3isp_preview_register_entities() 2234 ret = omap3isp_video_register(&prev->video_out, vdev); omap3isp_preview_register_entities() 2241 omap3isp_preview_unregister_entities(prev); omap3isp_preview_register_entities() 2251 * @prev : Pointer to preview structure 2254 static int preview_init_entities(struct isp_prev_device *prev) preview_init_entities() argument 2256 struct v4l2_subdev *sd = &prev->subdev; preview_init_entities() 2257 struct media_pad *pads = prev->pads; preview_init_entities() 2261 prev->input = PREVIEW_INPUT_NONE; preview_init_entities() 2267 v4l2_set_subdevdata(sd, prev); preview_init_entities() 2270 v4l2_ctrl_handler_init(&prev->ctrls, 2); preview_init_entities() 2271 v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_BRIGHTNESS, preview_init_entities() 2274 v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_CONTRAST, preview_init_entities() 2277 v4l2_ctrl_handler_setup(&prev->ctrls); preview_init_entities() 2278 sd->ctrl_handler = &prev->ctrls; preview_init_entities() 2295 prev->video_in.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; preview_init_entities() 2296 prev->video_in.ops = &preview_video_ops; preview_init_entities() 2297 prev->video_in.isp = to_isp_device(prev); preview_init_entities() 2298 prev->video_in.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; preview_init_entities() 2299 prev->video_in.bpl_alignment = 64; preview_init_entities() 2300 prev->video_out.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; preview_init_entities() 2301 prev->video_out.ops = &preview_video_ops; preview_init_entities() 2302 prev->video_out.isp = to_isp_device(prev); preview_init_entities() 2303 prev->video_out.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; preview_init_entities() 2304 prev->video_out.bpl_alignment = 32; preview_init_entities() 2306 ret = omap3isp_video_init(&prev->video_in, "preview"); preview_init_entities() 2310 ret = omap3isp_video_init(&prev->video_out, "preview"); preview_init_entities() 2315 ret = media_entity_create_link(&prev->video_in.video.entity, 0, preview_init_entities() 2316 &prev->subdev.entity, PREV_PAD_SINK, 0); preview_init_entities() 2320 ret = media_entity_create_link(&prev->subdev.entity, PREV_PAD_SOURCE, preview_init_entities() 2321 &prev->video_out.video.entity, 0, 0); preview_init_entities() 2328 omap3isp_video_cleanup(&prev->video_out); preview_init_entities() 2330 omap3isp_video_cleanup(&prev->video_in); preview_init_entities() 2332 media_entity_cleanup(&prev->subdev.entity); preview_init_entities() 2343 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_init() local 2345 init_waitqueue_head(&prev->wait); omap3isp_preview_init() 2347 preview_init_params(prev); omap3isp_preview_init() 2349 return preview_init_entities(prev); omap3isp_preview_init() 2354 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_cleanup() local 2356 v4l2_ctrl_handler_free(&prev->ctrls); omap3isp_preview_cleanup() 2357 omap3isp_video_cleanup(&prev->video_in); omap3isp_preview_cleanup() 2358 omap3isp_video_cleanup(&prev->video_out); omap3isp_preview_cleanup() 2359 media_entity_cleanup(&prev->subdev.entity); omap3isp_preview_cleanup()
|
/linux-4.4.14/arch/frv/include/asm/ |
H A D | switch_to.h | 18 * switch_to(prev, next) should switch from task `prev' to `next' 19 * `prev' will never be the same as `next'. 25 struct task_struct *prev); 27 #define switch_to(prev, next, last) \ 29 (prev)->thread.sched_lr = \ 31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
H A D | mmu_context.h | 35 #define switch_mm(prev, next, tsk) \ 37 if (prev != next) \ 38 change_mm_context(&prev->context, &next->context, next->pgd); \ 41 #define activate_mm(prev, next) \ 43 change_mm_context(&prev->context, &next->context, next->pgd); \
|
/linux-4.4.14/arch/cris/include/asm/ |
H A D | switch_to.h | 8 extern struct task_struct *resume(struct task_struct *prev, struct task_struct *next, int); 9 #define switch_to(prev,next,last) last = resume(prev,next, \
|
H A D | mmu_context.h | 9 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 14 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 19 switch_mm(prev, next, NULL); activate_mm()
|
H A D | fasttimer.h | 21 struct fast_timer *prev; member in struct:fast_timer
|
/linux-4.4.14/arch/mn10300/include/asm/ |
H A D | switch_to.h | 22 #define switch_fpu(prev, next) \ 24 if ((prev)->thread.fpu_flags & THREAD_HAS_FPU) { \ 25 (prev)->thread.fpu_flags &= ~THREAD_HAS_FPU; \ 26 (prev)->thread.uregs->epsw &= ~EPSW_FE; \ 27 fpu_save(&(prev)->thread.fpu_state); \ 31 #define switch_fpu(prev, next) do {} while (0) 36 struct task_struct *__switch_to(struct thread_struct *prev, 40 #define switch_to(prev, next, last) \ 42 switch_fpu(prev, next); \ 44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
H A D | mmu_context.h | 139 * @prev: The outgoing MM context. 143 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 148 if (prev != next) { switch_mm() 159 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.4.14/kernel/locking/ |
H A D | osq_lock.c | 38 struct optimistic_spin_node *prev) osq_wait_next() 45 * If there is a prev node in queue, then the 'old' value will be osq_wait_next() 46 * the prev node's CPU #, else it's set to OSQ_UNLOCKED_VAL since if osq_wait_next() 49 old = prev ? prev->cpu : OSQ_UNLOCKED_VAL; osq_wait_next() 55 * We were the last queued, we moved @lock back. @prev osq_wait_next() 65 * @node->next might complete Step-A and think its @prev is osq_wait_next() 87 struct optimistic_spin_node *prev, *next; osq_lock() local 105 prev = decode_cpu(old); osq_lock() 106 node->prev = prev; osq_lock() 107 WRITE_ONCE(prev->next, node); osq_lock() 110 * Normally @prev is untouchable after the above store; because at that osq_lock() 131 * Step - A -- stabilize @prev osq_lock() 133 * Undo our @prev->next assignment; this will make @prev's osq_lock() 139 if (prev->next == node && osq_lock() 140 cmpxchg(&prev->next, node, NULL) == node) osq_lock() 155 * case its step-C will write us a new @node->prev pointer. osq_lock() 157 prev = READ_ONCE(node->prev); osq_lock() 164 * back to @prev. osq_lock() 167 next = osq_wait_next(lock, node, prev); osq_lock() 174 * @prev is stable because its still waiting for a new @prev->next osq_lock() 179 WRITE_ONCE(next->prev, prev); osq_lock() 180 WRITE_ONCE(prev->next, next); osq_lock() 36 osq_wait_next(struct optimistic_spin_queue *lock, struct optimistic_spin_node *node, struct optimistic_spin_node *prev) osq_wait_next() argument
|
H A D | mcs_spinlock.h | 64 struct mcs_spinlock *prev; mcs_spin_lock() local 76 prev = xchg(lock, node); mcs_spin_lock() 77 if (likely(prev == NULL)) { mcs_spin_lock() 88 WRITE_ONCE(prev->next, node); mcs_spin_lock()
|
H A D | mutex.h | 17 __list_del((waiter)->list.prev, (waiter)->list.next)
|
/linux-4.4.14/arch/metag/include/asm/ |
H A D | mmu_context.h | 72 static inline void switch_mmu(struct mm_struct *prev, struct mm_struct *next) switch_mmu() argument 79 /* prev->context == prev->pgd in the case where we are initially switch_mmu() 81 if (prev->context.pgd_base != (unsigned long) prev->pgd) { switch_mmu() 83 ((pgd_t *) prev->context.pgd_base)[i] = prev->pgd[i]; switch_mmu() 85 prev->pgd = (pgd_t *)mmu_get_base(); switch_mmu() 87 next->pgd = prev->pgd; switch_mmu() 88 prev->pgd = (pgd_t *) prev->context.pgd_base; switch_mmu() 98 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 101 if (prev != next) switch_mm() 102 switch_mmu(prev, next); switch_mm()
|
/linux-4.4.14/arch/mips/include/asm/ |
H A D | switch_to.h | 25 * @prev: The task previously executed. 29 * This function is used whilst scheduling to save the context of prev & load 30 * the context of next. Returns prev. 32 extern asmlinkage struct task_struct *resume(struct task_struct *prev, 52 #define __mips_mt_fpaff_switch_to(prev) \ 54 struct thread_info *__prev_ti = task_thread_info(prev); \ 58 (!(KSTK_STATUS(prev) & ST0_CU1))) { \ 60 prev->cpus_allowed = prev->thread.user_cpus_allowed; \ 66 #define __mips_mt_fpaff_switch_to(prev) do { (void) (prev); } while (0) 84 #define switch_to(prev, next, last) \ 86 __mips_mt_fpaff_switch_to(prev); \ 87 lose_fpu_inatomic(1, prev); \ 89 __save_dsp(prev); \ 94 if ((KSTK_STATUS(prev) & ST0_CU2)) { \ 96 KSTK_STATUS(prev) &= ~ST0_CU2; \ 97 cop2_save(prev); \ 109 (last) = resume(prev, next, task_thread_info(next)); \
|
H A D | mmu_context.h | 139 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 157 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm() 179 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 194 cpumask_clear_cpu(cpu, mm_cpumask(prev)); activate_mm()
|
/linux-4.4.14/arch/avr32/include/asm/ |
H A D | switch_to.h | 18 #define ocd_switch(prev, next) \ 20 ocd_write(PID, prev->pid); \ 24 #define ocd_switch(prev, next) 28 * switch_to(prev, next, last) should switch from task `prev' to task 29 * `next'. `prev' will never be the same as `next'. 41 #define switch_to(prev, next, last) \ 43 ocd_switch(prev, next); \ 44 last = __switch_to(prev, &prev->thread.cpu_context + 1, \
|
H A D | mmu_context.h | 108 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument 112 if (likely(prev != next)) { switch_mm() 122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.4.14/drivers/fmc/ |
H A D | fmc-dump.c | 25 const uint8_t *prev) dump_line() 29 if (!prev || memcmp(line, prev, LINELEN)) { dump_line() 39 if (line == prev + LINELEN) dump_line() 41 return prev; dump_line() 46 const uint8_t *line, *prev; fmc_dump_eeprom() local 58 prev = NULL; fmc_dump_eeprom() 60 prev = dump_line(i, line, prev); fmc_dump_eeprom() 65 const uint8_t *line, *prev; fmc_dump_sdb() local 96 prev = NULL; fmc_dump_sdb() 98 prev = dump_line(i, line, prev); fmc_dump_sdb() 24 dump_line(int addr, const uint8_t *line, const uint8_t *prev) dump_line() argument
|
/linux-4.4.14/arch/powerpc/include/asm/ |
H A D | cmpxchg.h | 18 unsigned long prev; __xchg_u32() local 27 : "=&r" (prev), "+m" (*(volatile unsigned int *)p) __xchg_u32() 31 return prev; __xchg_u32() 43 unsigned long prev; __xchg_u32_local() local 50 : "=&r" (prev), "+m" (*(volatile unsigned int *)p) __xchg_u32_local() 54 return prev; __xchg_u32_local() 61 unsigned long prev; __xchg_u64() local 70 : "=&r" (prev), "+m" (*(volatile unsigned long *)p) __xchg_u64() 74 return prev; __xchg_u64() 80 unsigned long prev; __xchg_u64_local() local 87 : "=&r" (prev), "+m" (*(volatile unsigned long *)p) __xchg_u64_local() 91 return prev; __xchg_u64_local() 151 unsigned int prev; __cmpxchg_u32() local 164 : "=&r" (prev), "+m" (*p) __cmpxchg_u32() 168 return prev; __cmpxchg_u32() 175 unsigned int prev; __cmpxchg_u32_local() local 186 : "=&r" (prev), "+m" (*p) __cmpxchg_u32_local() 190 return prev; __cmpxchg_u32_local() 197 unsigned long prev; __cmpxchg_u64() local 209 : "=&r" (prev), "+m" (*p) __cmpxchg_u64() 213 return prev; __cmpxchg_u64() 220 unsigned long prev; __cmpxchg_u64_local() local 230 : "=&r" (prev), "+m" (*p) __cmpxchg_u64_local() 234 return prev; __cmpxchg_u64_local()
|
H A D | switch_to.h | 13 #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next))) 16 extern struct task_struct *_switch(struct thread_struct *prev, 19 static inline void save_early_sprs(struct thread_struct *prev) save_early_sprs() argument 22 prev->tar = mfspr(SPRN_TAR); save_early_sprs() 24 prev->dscr = mfspr(SPRN_DSCR); save_early_sprs() 27 static inline void save_early_sprs(struct thread_struct *prev) {} save_early_sprs() argument
|
H A D | mmu_context.h | 37 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next); 59 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 75 if (prev == next) switch_mm() 79 /* Switch coprocessor context only if prev or next uses a coprocessor */ switch_mm() 80 if (prev->context.acop || next->context.acop) switch_mm() 99 switch_mmu_context(prev, next); switch_mm() 110 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 115 switch_mm(prev, next, current); activate_mm()
|
H A D | futex.h | 90 u32 prev; futex_atomic_cmpxchg_inatomic() local 112 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic() 116 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.4.14/arch/arm/include/asm/ |
H A D | switch_to.h | 19 * switch_to(prev, next) should switch from task `prev' to `next' 20 * `prev' will never be the same as `next'. schedule() itself 25 #define switch_to(prev,next,last) \ 28 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
|
H A D | mmu_context.h | 93 #define activate_mm(prev,next) switch_mm(prev, next, NULL) 116 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 132 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { switch_mm() 135 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
|
/linux-4.4.14/arch/um/os-Linux/ |
H A D | irq.c | 82 struct irq_fd **prev; os_free_irq_by_cb() local 85 prev = &active_fds; os_free_irq_by_cb() 86 while (*prev != NULL) { os_free_irq_by_cb() 87 if ((*test)(*prev, arg)) { os_free_irq_by_cb() 88 struct irq_fd *old_fd = *prev; os_free_irq_by_cb() 90 (pollfds[i].fd != (*prev)->fd)) { os_free_irq_by_cb() 94 (*prev)->fd, pollfds[i].fd); os_free_irq_by_cb() 107 *last_irq_ptr2 = prev; os_free_irq_by_cb() 109 *prev = (*prev)->next; os_free_irq_by_cb() 115 prev = &(*prev)->next; os_free_irq_by_cb()
|
/linux-4.4.14/arch/microblaze/include/asm/ |
H A D | switch_to.h | 15 extern struct task_struct *_switch_to(struct thread_info *prev, 18 #define switch_to(prev, next, last) \ 20 (last) = _switch_to(task_thread_info(prev), \
|
H A D | futex.h | 101 u32 prev; futex_atomic_cmpxchg_inatomic() local 120 : "+r" (ret), "=&r" (prev), "=&r"(cmp) \ futex_atomic_cmpxchg_inatomic() 123 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | switch_to.h | 44 #define __switch_to(prev,next,last) do { \ 45 if (IA64_HAS_EXTRA_STATE(prev)) \ 46 ia64_save_extra(prev); \ 60 # define switch_to(prev,next,last) do { \ 61 if (ia64_psr(task_pt_regs(prev))->mfh && ia64_is_local_fpu_owner(prev)) { \ 62 ia64_psr(task_pt_regs(prev))->mfh = 0; \ 63 (prev)->thread.flags |= IA64_THREAD_FPH_VALID; \ 64 __ia64_save_fpu((prev)->thread.fph); \ 66 __switch_to(prev, next, last); \ 76 # define switch_to(prev,next,last) __switch_to(prev, next, last)
|
H A D | futex.h | 110 unsigned long prev; futex_atomic_cmpxchg_inatomic() local 117 : "+r" (r8), "=&r" (prev) futex_atomic_cmpxchg_inatomic() 121 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.4.14/arch/c6x/include/asm/ |
H A D | switch_to.h | 20 asmlinkage void *__switch_to(struct thread_struct *prev, 24 #define switch_to(prev, next, last) \ 27 (last) = __switch_to(&(prev)->thread, \ 28 &(next)->thread, (prev)); \
|
/linux-4.4.14/include/linux/ |
H A D | user-return-notifier.h | 18 static inline void propagate_user_return_notify(struct task_struct *prev, propagate_user_return_notify() argument 21 if (test_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY)) { propagate_user_return_notify() 22 clear_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY); propagate_user_return_notify() 38 static inline void propagate_user_return_notify(struct task_struct *prev, propagate_user_return_notify() argument
|
H A D | vtime.h | 42 extern void vtime_task_switch(struct task_struct *prev); vtime_accounting_enabled() 44 extern void vtime_common_task_switch(struct task_struct *prev); vtime_task_switch() 45 static inline void vtime_task_switch(struct task_struct *prev) vtime_task_switch() argument 48 vtime_common_task_switch(prev); vtime_task_switch() 69 static inline void vtime_task_switch(struct task_struct *prev) { } vtime_account_system() argument
|
H A D | list.h | 15 * sometimes we already know the next/prev entries and we can 28 list->prev = list; INIT_LIST_HEAD() 35 * the prev/next entries already! 39 struct list_head *prev, __list_add() 42 next->prev = new; __list_add() 44 new->prev = prev; __list_add() 45 prev->next = new; __list_add() 49 struct list_head *prev, 77 __list_add(new, head->prev, head); list_add_tail() 81 * Delete a list entry by making the prev/next entries 85 * the prev/next entries already! 87 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument 89 next->prev = prev; __list_del() 90 WRITE_ONCE(prev->next, next); __list_del() 102 __list_del(entry->prev, entry->next); __list_del_entry() 107 __list_del(entry->prev, entry->next); list_del() 109 entry->prev = LIST_POISON2; list_del() 127 new->next->prev = new; list_replace() 128 new->prev = old->prev; list_replace() 129 new->prev->next = new; list_replace() 198 * in the process of modifying either member (next or prev) 208 return (next == head) && (next == head->prev); list_empty_careful() 231 return !list_empty(head) && (head->next == head->prev); list_is_singular() 239 list->next->prev = list; __list_cut_position() 240 list->prev = entry; __list_cut_position() 243 new_first->prev = head; __list_cut_position() 275 struct list_head *prev, __list_splice() 279 struct list_head *last = list->prev; __list_splice() 281 first->prev = prev; __list_splice() 282 prev->next = first; __list_splice() 285 next->prev = last; __list_splice() 309 __list_splice(list, head->prev, head); list_splice_tail() 340 __list_splice(list, head->prev, head); list_splice_tail_init() 374 list_entry((ptr)->prev, type, member) 396 * list_prev_entry - get the prev element in list 401 list_entry((pos)->member.prev, typeof(*(pos)), member) 417 for (pos = (head)->prev; pos != (head); pos = pos->prev) 436 for (pos = (head)->prev, n = pos->prev; \ 438 pos = n, n = pos->prev) 660 struct hlist_node *prev) hlist_add_behind() 662 n->next = prev->next; hlist_add_behind() 663 prev->next = n; hlist_add_behind() 664 n->pprev = &prev->next; hlist_add_behind() 38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument 659 hlist_add_behind(struct hlist_node *n, struct hlist_node *prev) hlist_add_behind() argument
|
H A D | rculist.h | 33 WRITE_ONCE(list->prev, list); INIT_LIST_HEAD_RCU() 46 * the prev/next entries already! 50 struct list_head *prev, struct list_head *next) __list_add_rcu() 53 new->prev = prev; __list_add_rcu() 54 rcu_assign_pointer(list_next_rcu(prev), new); __list_add_rcu() 55 next->prev = new; __list_add_rcu() 59 struct list_head *prev, struct list_head *next); 102 __list_add_rcu(new, head->prev, head); list_add_tail_rcu() 132 entry->prev = LIST_POISON2; list_del_rcu() 175 new->prev = old->prev; list_replace_rcu() 176 rcu_assign_pointer(list_next_rcu(new->prev), new); list_replace_rcu() 177 new->next->prev = new; list_replace_rcu() 178 old->prev = LIST_POISON2; list_replace_rcu() 203 struct list_head *last = list->prev; list_splice_init_rcu() 230 * permitted to traverse the prev pointers without excluding list_splice_init_rcu() 236 first->prev = head; list_splice_init_rcu() 237 at->prev = last; list_splice_init_rcu() 434 * @prev: the existing element to add the new element after. 450 struct hlist_node *prev) hlist_add_behind_rcu() 452 n->next = prev->next; hlist_add_behind_rcu() 453 n->pprev = &prev->next; hlist_add_behind_rcu() 454 rcu_assign_pointer(hlist_next_rcu(prev), n); hlist_add_behind_rcu() 49 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument 449 hlist_add_behind_rcu(struct hlist_node *n, struct hlist_node *prev) hlist_add_behind_rcu() argument
|
H A D | osq_lock.h | 9 struct optimistic_spin_node *next, *prev; member in struct:optimistic_spin_node
|
H A D | of_pdt.h | 20 * If prev is NULL, the first property will be returned. 22 int (*nextprop)(phandle node, char *prev, char *buf);
|
H A D | agpgart.h | 93 struct agp_client *prev; member in struct:agp_client 101 struct agp_controller *prev; member in struct:agp_controller 116 struct agp_file_private *prev; member in struct:agp_file_private
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | switch_to_64.h | 19 struct task_struct *sh64_switch_to(struct task_struct *prev, 24 #define switch_to(prev,next,last) \ 30 last = sh64_switch_to(prev, &prev->thread, next, \
|
H A D | futex-irq.h | 96 u32 prev = 0; atomic_futex_op_cmpxchg_inatomic() local 100 ret = get_user(prev, uaddr); atomic_futex_op_cmpxchg_inatomic() 101 if (!ret && oldval == prev) atomic_futex_op_cmpxchg_inatomic() 106 *uval = prev; atomic_futex_op_cmpxchg_inatomic()
|
H A D | switch_to_32.h | 63 struct task_struct *__switch_to(struct task_struct *prev, 69 #define switch_to(prev, next, last) \ 79 if (is_dsp_enabled(prev)) \ 80 __save_dsp(prev); \ 84 __ts1 = (u32 *)&prev->thread.sp; \ 85 __ts2 = (u32 *)&prev->thread.pc; \ 86 __ts4 = (u32 *)prev; \
|
H A D | mmu_context.h | 118 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument 124 if (likely(prev != next)) { switch_mm() 133 #define activate_mm(prev, next) switch_mm((prev),(next),NULL)
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | switch_to.h | 13 * switch_to(prev, next) should switch from task `prev' to `next' 14 * `prev' will never be the same as `next'. 16 * `next' and `prev' should be struct task_struct, but it isn't always defined 28 #define switch_to(prev, next, last) do { \ 37 " push %1 ; store `prev' on new stack \n" \ 44 : "0" (prev), \ 45 "r" (&(prev->thread.sp)), "r" (&(next->thread.sp)), \ 46 "r" (&(prev->thread.lr)), "r" (&(next->thread.lr)) \
|
H A D | mmu_context.h | 121 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument 128 if (prev != next) { switch_mm() 145 #define activate_mm(prev, next) \ 146 switch_mm((prev), (next), NULL) 155 #define switch_mm(prev,next,tsk) do { } while (0) 157 #define activate_mm(prev,next) do { } while (0)
|
/linux-4.4.14/tools/usb/usbip/libsrc/ |
H A D | list.h | 13 * sometimes we already know the next/prev entries and we can 19 struct list_head *next, *prev; member in struct:list_head 30 list->prev = list; INIT_LIST_HEAD() 37 * the prev/next entries already! 40 struct list_head *prev, __list_add() 43 next->prev = new; __list_add() 45 new->prev = prev; __list_add() 46 prev->next = new; __list_add() 63 * Delete a list entry by making the prev/next entries 67 * the prev/next entries already! 69 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument 71 next->prev = prev; __list_del() 72 prev->next = next; __list_del() 87 __list_del(entry->prev, entry->next); __list_del_entry() 92 __list_del(entry->prev, entry->next); list_del() 94 entry->prev = LIST_POISON2; list_del() 39 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.4.14/arch/arc/include/asm/ |
H A D | switch_to.h | 31 #define switch_to(prev, next, last) \ 33 ARC_FPU_PREV(prev, next); \ 34 last = __switch_to(prev, next);\
|
H A D | cmpxchg.h | 22 unsigned long prev; __cmpxchg() local 36 : "=&r"(prev) /* Early clobber, to prevent reg reuse */ __cmpxchg() 44 return prev; __cmpxchg() 53 int prev; __cmpxchg() local 60 prev = *p; __cmpxchg() 61 if (prev == expected) __cmpxchg() 64 return prev; __cmpxchg()
|
/linux-4.4.14/scripts/kconfig/ |
H A D | list.h | 24 struct list_head *next, *prev; member in struct:list_head 79 * the prev/next entries already! 82 struct list_head *prev, __list_add() 85 next->prev = _new; __list_add() 87 _new->prev = prev; __list_add() 88 prev->next = _new; __list_add() 101 __list_add(_new, head->prev, head); list_add_tail() 105 * Delete a list entry by making the prev/next entries 109 * the prev/next entries already! 111 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 113 next->prev = prev; __list_del() 114 prev->next = next; __list_del() 127 __list_del(entry->prev, entry->next); list_del() 129 entry->prev = (struct list_head*)LIST_POISON2; list_del() 81 __list_add(struct list_head *_new, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | cmpxchg_32.h | 27 u64 prev = *ptr; set_64bit() local 32 : "=m" (*ptr), "+A" (prev) set_64bit() 48 u64 prev; __cmpxchg64() local 50 : "=A" (prev), __cmpxchg64() 56 return prev; __cmpxchg64() 61 u64 prev; __cmpxchg64_local() local 63 : "=A" (prev), __cmpxchg64_local() 69 return prev; __cmpxchg64_local()
|
H A D | switch_to.h | 5 __visible struct task_struct *__switch_to(struct task_struct *prev, 31 #define switch_to(prev, next, last) \ 55 : [prev_sp] "=m" (prev->thread.sp), \ 56 [prev_ip] "=m" (prev->thread.ip), \ 70 [prev] "a" (prev), \ 108 #define switch_to(prev, next, last) \ 110 "movq %%rsp,%P[threadrsp](%[prev])\n\t" /* save RSP */ \ 122 : [next] "S" (next), [prev] "D" (prev), \
|
H A D | mmu_context.h | 15 static inline void paravirt_activate_mm(struct mm_struct *prev, paravirt_activate_mm() argument 107 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 112 if (likely(prev != next)) { switch_mm() 152 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm() 161 * It's possible that prev->context.ldt doesn't match switch_mm() 162 * the LDT register. This can happen if leave_mm(prev) switch_mm() 164 * prev->context.ldt but suppressed an IPI to this CPU. switch_mm() 165 * In this case, prev->context.ldt != NULL, because we switch_mm() 168 * prev->context.ldt, because mms never share an LDT. switch_mm() 170 if (unlikely(prev->context.ldt != next->context.ldt)) switch_mm() 205 #define activate_mm(prev, next) \ 207 paravirt_activate_mm((prev), (next)); \ 208 switch_mm((prev), (next), NULL); \
|
/linux-4.4.14/arch/xtensa/include/asm/ |
H A D | nommu_context.h | 18 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 22 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
|
H A D | switch_to.h | 17 #define switch_to(prev,next,last) \ 19 (last) = _switch_to(prev, next); \
|
H A D | futex.h | 112 u32 prev; futex_atomic_cmpxchg_inatomic() local 138 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic() 142 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
H A D | mmu_context.h | 123 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 133 if (migrated || prev != next) switch_mm() 137 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.4.14/scripts/gdb/linux/ |
H A D | lists.py | 34 p = c['prev'].dereference() 38 gdb.write('prev.next != current: ' 40 'prev@{p_addr}={p}\n'.format( 48 gdb.write('prev is not accessible: ' 55 if n['prev'] != c.address: 56 gdb.write('next.prev != current: '
|
/linux-4.4.14/arch/parisc/lib/ |
H A D | bitops.c | 62 unsigned long prev; __cmpxchg_u64() local 65 if ((prev = *ptr) == old) __cmpxchg_u64() 68 return prev; __cmpxchg_u64() 75 unsigned int prev; __cmpxchg_u32() local 78 if ((prev = *ptr) == old) __cmpxchg_u32() 81 return (unsigned long)prev; __cmpxchg_u32()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | mm.c | 53 struct nvkm_mm_node *prev = node(this, prev); nvkm_mm_free() local 56 if (prev && prev->type == NVKM_MM_TYPE_NONE) { nvkm_mm_free() 57 prev->length += this->length; nvkm_mm_free() 59 kfree(this); this = prev; nvkm_mm_free() 72 list_for_each_entry(prev, &mm->free, fl_entry) { nvkm_mm_free() 73 if (this->offset < prev->offset) nvkm_mm_free() 77 list_add_tail(&this->fl_entry, &prev->fl_entry); nvkm_mm_free() 114 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_head() local 129 prev = node(this, prev); nvkm_mm_head() 130 if (prev && prev->type != type) nvkm_mm_head() 188 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_tail() local 202 prev = node(this, prev); nvkm_mm_tail() 203 if (prev && prev->type != type) nvkm_mm_tail() 240 struct nvkm_mm_node *node, *prev; nvkm_mm_init() local 244 prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); nvkm_mm_init() 245 next = prev->offset + prev->length; nvkm_mm_init()
|
/linux-4.4.14/drivers/net/fddi/skfp/ |
H A D | smttimer.c | 41 struct smt_timer **prev ; smt_timer_stop() local 51 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { smt_timer_stop() 53 *prev = tm->tm_next ; smt_timer_stop() 65 struct smt_timer **prev ; smt_timer_start() local 92 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { smt_timer_start() 99 *prev = timer ; smt_timer_start()
|
/linux-4.4.14/net/bridge/ |
H A D | br_forward.c | 24 static int deliver_clone(const struct net_bridge_port *prev, 156 static int deliver_clone(const struct net_bridge_port *prev, deliver_clone() argument 169 __packet_hook(prev, skb); deliver_clone() 174 struct net_bridge_port *prev, struct net_bridge_port *p, maybe_deliver() 182 return prev; maybe_deliver() 184 if (!prev) maybe_deliver() 187 err = deliver_clone(prev, skb, __packet_hook); maybe_deliver() 203 struct net_bridge_port *prev; br_flood() local 205 prev = NULL; br_flood() 219 prev = maybe_deliver(prev, p, skb, __packet_hook); br_flood() 220 if (IS_ERR(prev)) br_flood() 224 if (!prev) br_flood() 228 deliver_clone(prev, skb, __packet_hook); br_flood() 230 __packet_hook(prev, skb); br_flood() 262 struct net_bridge_port *prev = NULL; br_multicast_flood() local 278 prev = maybe_deliver(prev, port, skb, __packet_hook); br_multicast_flood() 279 if (IS_ERR(prev)) br_multicast_flood() 288 if (!prev) br_multicast_flood() 292 deliver_clone(prev, skb, __packet_hook); br_multicast_flood() 294 __packet_hook(prev, skb); br_multicast_flood() 173 maybe_deliver( struct net_bridge_port *prev, struct net_bridge_port *p, struct sk_buff *skb, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) maybe_deliver() argument
|
/linux-4.4.14/drivers/gpu/drm/radeon/ |
H A D | mkregtable.c | 36 * sometimes we already know the next/prev entries and we can 42 struct list_head *next, *prev; member in struct:list_head 53 list->prev = list; INIT_LIST_HEAD() 60 * the prev/next entries already! 64 struct list_head *prev, struct list_head *next) __list_add() 66 next->prev = new; __list_add() 68 new->prev = prev; __list_add() 69 prev->next = new; __list_add() 73 struct list_head *prev, struct list_head *next); 99 __list_add(new, head->prev, head); list_add_tail() 103 * Delete a list entry by making the prev/next entries 107 * the prev/next entries already! 109 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 111 next->prev = prev; __list_del() 112 prev->next = next; __list_del() 124 __list_del(entry->prev, entry->next); list_del() 126 entry->prev = (void *)0xBEEFDEAD; list_del() 142 new->next->prev = new; list_replace() 143 new->prev = old->prev; list_replace() 144 new->prev->next = new; list_replace() 160 __list_del(entry->prev, entry->next); list_del_init() 171 __list_del(list->prev, list->next); list_move() 183 __list_del(list->prev, list->next); list_move_tail() 213 * in the process of modifying either member (next or prev) 223 return (next == head) && (next == head->prev); list_empty_careful() 232 return !list_empty(head) && (head->next == head->prev); list_is_singular() 241 list->next->prev = list; __list_cut_position() 242 list->prev = entry; __list_cut_position() 245 new_first->prev = head; __list_cut_position() 277 struct list_head *prev, struct list_head *next) __list_splice() 280 struct list_head *last = list->prev; __list_splice() 282 first->prev = prev; __list_splice() 283 prev->next = first; __list_splice() 286 next->prev = last; __list_splice() 310 __list_splice(list, head->prev, head); list_splice_tail() 341 __list_splice(list, head->prev, head); list_splice_tail_init() 381 for (pos = (head)->prev; prefetch(pos->prev), pos != (head); \ 382 pos = pos->prev) 401 for (pos = (head)->prev, n = pos->prev; \ 402 prefetch(pos->prev), pos != (head); \ 403 pos = n, n = pos->prev) 423 for (pos = list_entry((head)->prev, typeof(*pos), member); \ 424 prefetch(pos->member.prev), &pos->member != (head); \ 425 pos = list_entry(pos->member.prev, typeof(*pos), member)) 462 for (pos = list_entry(pos->member.prev, typeof(*pos), member); \ 463 prefetch(pos->member.prev), &pos->member != (head); \ 464 pos = list_entry(pos->member.prev, typeof(*pos), member)) 533 for (pos = list_entry((head)->prev, typeof(*pos), member), \ 534 n = list_entry(pos->member.prev, typeof(*pos), member); \ 536 pos = n, n = list_entry(n->member.prev, typeof(*n), member)) 63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 276 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
|
H A D | radeon_mem.c | 55 newblock->prev = p; split_block() 56 p->next->prev = newblock; split_block() 72 newblock->prev = p; split_block() 73 p->next->prev = newblock; split_block() 121 p->next->prev = p; free_block() 125 if (p->prev->file_priv == NULL) { free_block() 126 struct mem_block *q = p->prev; free_block() 129 q->next->prev = q; free_block() 152 blocks->next = blocks->prev = *heap; init_heap() 155 (*heap)->next = (*heap)->prev = blocks; init_heap() 181 p->next->prev = p; list_for_each()
|
/linux-4.4.14/arch/cris/arch-v32/mm/ |
H A D | intmem.c | 107 struct intmem_allocation *prev = crisv32_intmem_free() local 108 list_entry(allocation->entry.prev, crisv32_intmem_free() 115 /* Join with prev and/or next if also free */ crisv32_intmem_free() 116 if ((prev != &intmem_allocations) && crisv32_intmem_free() 117 (prev->status == STATUS_FREE)) { crisv32_intmem_free() 118 prev->size += allocation->size; crisv32_intmem_free() 121 allocation = prev; crisv32_intmem_free()
|
/linux-4.4.14/mm/ |
H A D | interval_tree.c | 28 /* Insert node immediately after prev in the interval tree */ vma_interval_tree_insert_after() 30 struct vm_area_struct *prev, vma_interval_tree_insert_after() 37 VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); vma_interval_tree_insert_after() 39 if (!prev->shared.rb.rb_right) { vma_interval_tree_insert_after() 40 parent = prev; vma_interval_tree_insert_after() 41 link = &prev->shared.rb.rb_right; vma_interval_tree_insert_after() 43 parent = rb_entry(prev->shared.rb.rb_right, vma_interval_tree_insert_after() 29 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root *root) vma_interval_tree_insert_after() argument
|
H A D | madvise.c | 47 struct vm_area_struct **prev, madvise_behavior() 100 *prev = vma; madvise_behavior() 105 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior() 108 if (*prev) { madvise_behavior() 109 vma = *prev; madvise_behavior() 113 *prev = vma; madvise_behavior() 221 struct vm_area_struct **prev, madvise_willneed() 228 *prev = vma; madvise_willneed() 234 *prev = vma; madvise_willneed() 249 *prev = vma; madvise_willneed() 279 struct vm_area_struct **prev, madvise_dontneed() 282 *prev = vma; madvise_dontneed() 295 struct vm_area_struct **prev, madvise_remove() 302 *prev = NULL; /* tell sys_madvise we drop mmap_sem */ madvise_remove() 374 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, madvise_vma() argument 379 return madvise_remove(vma, prev, start, end); madvise_vma() 381 return madvise_willneed(vma, prev, start, end); madvise_vma() 383 return madvise_dontneed(vma, prev, start, end); madvise_vma() 385 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma() 463 struct vm_area_struct *vma, *prev; SYSCALL_DEFINE3() local 504 vma = find_vma_prev(current->mm, start, &prev); SYSCALL_DEFINE3() 506 prev = vma; SYSCALL_DEFINE3() 529 error = madvise_vma(vma, &prev, start, tmp, behavior); SYSCALL_DEFINE3() 533 if (prev && start < prev->vm_end) SYSCALL_DEFINE3() 534 start = prev->vm_end; SYSCALL_DEFINE3() 538 if (prev) SYSCALL_DEFINE3() 539 vma = prev->vm_next; SYSCALL_DEFINE3() 46 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument 220 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument 278 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument 294 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument
|
H A D | slob.c | 219 slob_t *prev, *cur, *aligned = NULL; slob_page_alloc() local 222 for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) { slob_page_alloc() 236 prev = cur; slob_page_alloc() 243 if (prev) slob_page_alloc() 244 set_slob(prev, slob_units(prev), next); slob_page_alloc() 248 if (prev) slob_page_alloc() 249 set_slob(prev, slob_units(prev), cur + units); slob_page_alloc() 271 struct list_head *prev; slob_alloc() local 299 prev = sp->lru.prev; list_for_each_entry() 307 if (prev != slob_list->prev && list_for_each_entry() 308 slob_list->next != prev->next) list_for_each_entry() 309 list_move_tail(slob_list, prev->next); list_for_each_entry() 343 slob_t *prev, *next, *b = (slob_t *)block; slob_free() local 399 prev = sp->freelist; slob_free() 400 next = slob_next(prev); slob_free() 402 prev = next; slob_free() 403 next = slob_next(prev); slob_free() 406 if (!slob_last(prev) && b + units == next) { slob_free() 412 if (prev + slob_units(prev) == b) { slob_free() 413 units = slob_units(b) + slob_units(prev); slob_free() 414 set_slob(prev, units, slob_next(b)); slob_free() 416 set_slob(prev, slob_units(prev), b); slob_free()
|
H A D | mlock.c | 498 static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev, mlock_fixup() argument 513 *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma, mlock_fixup() 516 if (*prev) { mlock_fixup() 517 vma = *prev; mlock_fixup() 554 *prev = vma; mlock_fixup() 562 struct vm_area_struct * vma, * prev; apply_vma_lock_flags() local 576 prev = vma->vm_prev; apply_vma_lock_flags() 578 prev = vma; apply_vma_lock_flags() 589 error = mlock_fixup(vma, &prev, nstart, tmp, newflags); apply_vma_lock_flags() 593 if (nstart < prev->vm_end) apply_vma_lock_flags() 594 nstart = prev->vm_end; apply_vma_lock_flags() 598 vma = prev->vm_next; apply_vma_lock_flags() 687 struct vm_area_struct * vma, * prev = NULL; apply_mlockall_flags() local 707 for (vma = current->mm->mmap; vma ; vma = prev->vm_next) { apply_mlockall_flags() 714 mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags); apply_mlockall_flags()
|
H A D | mmap.c | 62 struct vm_area_struct *vma, struct vm_area_struct *prev, 382 unsigned long prev = 0, pend = 0; browse_rb() local 387 if (vma->vm_start < prev) { browse_rb() 388 pr_emerg("vm_start %lx < prev %lx\n", browse_rb() 389 vma->vm_start, prev); browse_rb() 410 prev = vma->vm_start; browse_rb() 661 struct vm_area_struct *prev, struct rb_node **rb_link, __vma_link() 664 __vma_link_list(mm, vma, prev, rb_parent); __vma_link() 669 struct vm_area_struct *prev, struct rb_node **rb_link, vma_link() 679 __vma_link(mm, vma, prev, rb_link, rb_parent); vma_link() 695 struct vm_area_struct *prev; __insert_vm_struct() local 699 &prev, &rb_link, &rb_parent)) __insert_vm_struct() 701 __vma_link(mm, vma, prev, rb_link, rb_parent); __insert_vm_struct() 707 struct vm_area_struct *prev) __vma_unlink() 712 prev->vm_next = next = vma->vm_next; __vma_unlink() 714 next->vm_prev = prev; __vma_unlink() 1019 * an offset within prev, or at the start of next), and the flags of 1025 * vma, PPPPPP is the prev vma specified, and NNNNNN the next vma after: 1041 struct vm_area_struct *prev, unsigned long addr, vma_merge() 1058 if (prev) vma_merge() 1059 next = prev->vm_next; vma_merge() 1069 if (prev && prev->vm_end == addr && vma_merge() 1070 mpol_equal(vma_policy(prev), policy) && vma_merge() 1071 can_vma_merge_after(prev, vm_flags, vma_merge() 1083 is_mergeable_anon_vma(prev->anon_vma, vma_merge() 1086 err = vma_adjust(prev, prev->vm_start, vma_merge() 1087 next->vm_end, prev->vm_pgoff, NULL); vma_merge() 1089 err = vma_adjust(prev, prev->vm_start, vma_merge() 1090 end, prev->vm_pgoff, NULL); vma_merge() 1093 khugepaged_enter_vma_merge(prev, vm_flags); vma_merge() 1094 return prev; vma_merge() 1105 if (prev && addr < prev->vm_end) /* case 4 */ vma_merge() 1106 err = vma_adjust(prev, prev->vm_start, vma_merge() 1107 addr, prev->vm_pgoff, NULL); vma_merge() 1545 struct vm_area_struct *vma, *prev; mmap_region() local 1568 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, mmap_region() 1587 vma = vma_merge(mm, prev, addr, addr + len, vm_flags, mmap_region() 1637 * Bug: If addr is changed, prev, rb_link, rb_parent should mmap_region() 1650 vma_link(mm, vma, prev, rb_link, rb_parent); mmap_region() 1692 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); mmap_region() 1776 struct rb_node *prev = &vma->vm_rb; unmapped_area() local 1777 if (!rb_parent(prev)) unmapped_area() 1779 vma = rb_entry(rb_parent(prev), unmapped_area() 1781 if (prev == vma->vm_rb.rb_left) { unmapped_area() 1879 struct rb_node *prev = &vma->vm_rb; unmapped_area_topdown() local 1880 if (!rb_parent(prev)) unmapped_area_topdown() 1882 vma = rb_entry(rb_parent(prev), unmapped_area_topdown() 1884 if (prev == vma->vm_rb.rb_right) { unmapped_area_topdown() 2320 struct vm_area_struct *vma, *prev; find_extend_vma() local 2323 vma = find_vma_prev(mm, addr, &prev); find_extend_vma() 2326 if (!prev || expand_stack(prev, addr)) find_extend_vma() 2328 if (prev->vm_flags & VM_LOCKED) find_extend_vma() 2329 populate_vma_page_range(prev, addr, prev->vm_end, NULL); find_extend_vma() 2330 return prev; find_extend_vma() 2335 struct vm_area_struct *prev; expand_stack() local 2338 prev = vma->vm_prev; expand_stack() 2339 if (prev && prev->vm_end == address) { expand_stack() 2340 if (!(prev->vm_flags & VM_GROWSDOWN)) expand_stack() 2401 struct vm_area_struct *vma, struct vm_area_struct *prev, unmap_region() 2404 struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; unmap_region() 2411 free_pgtables(&tlb, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, unmap_region() 2422 struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped() 2427 insertion_point = (prev ? &prev->vm_next : &mm->mmap); detach_vmas_to_be_unmapped() 2437 vma->vm_prev = prev; detach_vmas_to_be_unmapped() 2440 mm->highest_vm_end = prev ? prev->vm_end : 0; detach_vmas_to_be_unmapped() 2535 struct vm_area_struct *vma, *prev, *last; do_munmap() local 2548 prev = vma->vm_prev; do_munmap() 2577 prev = vma; do_munmap() 2587 vma = prev ? prev->vm_next : mm->mmap; do_munmap() 2606 detach_vmas_to_be_unmapped(mm, vma, prev, end); do_munmap() 2607 unmap_region(mm, vma, prev, start, end); do_munmap() 2746 struct vm_area_struct *vma, *prev; do_brk() local 2775 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, do_brk() 2792 vma = vma_merge(mm, prev, addr, addr + len, flags, do_brk() 2813 vma_link(mm, vma, prev, rb_link, rb_parent); do_brk() 2892 struct vm_area_struct *prev; insert_vm_struct() local 2896 &prev, &rb_link, &rb_parent)) insert_vm_struct() 2919 vma_link(mm, vma, prev, rb_link, rb_parent); insert_vm_struct() 2934 struct vm_area_struct *new_vma, *prev; copy_vma() local 2947 if (find_vma_links(mm, addr, addr + len, &prev, &rb_link, &rb_parent)) copy_vma() 2949 new_vma = vma_merge(mm, prev, addr, addr + len, vma->vm_flags, copy_vma() 2991 vma_link(mm, new_vma, prev, rb_link, rb_parent); copy_vma() 660 __vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) __vma_link() argument 668 vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) vma_link() argument 706 __vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev) __vma_unlink() argument 1040 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy, struct vm_userfaultfd_ctx vm_userfaultfd_ctx) vma_merge() argument 2400 unmap_region(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long start, unsigned long end) unmap_region() argument 2421 detach_vmas_to_be_unmapped(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped() argument
|
H A D | mprotect.c | 351 struct vm_area_struct *vma, *prev; SYSCALL_DEFINE3() local 384 prev = vma->vm_prev; SYSCALL_DEFINE3() 403 prev = vma; SYSCALL_DEFINE3() 426 error = mprotect_fixup(vma, &prev, nstart, tmp, newflags); SYSCALL_DEFINE3() 431 if (nstart < prev->vm_end) SYSCALL_DEFINE3() 432 nstart = prev->vm_end; SYSCALL_DEFINE3() 436 vma = prev->vm_next; SYSCALL_DEFINE3()
|
/linux-4.4.14/sound/pci/ctxfi/ |
H A D | ctimap.c | 46 pre = pos->prev; 48 pre = head->prev; 50 __list_add(&entry->list, pos->prev, pos); 52 pre = head->prev; 79 pre = (entry->list.prev == head) ? head->prev : entry->list.prev; input_mapper_delete()
|
/linux-4.4.14/arch/alpha/include/asm/ |
H A D | xchg.h | 132 * prev is equal to old) then we aren't acquiring anything new and 139 unsigned long prev, tmp, cmp, addr64; ____cmpxchg() local 157 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) ____cmpxchg() 160 return prev; ____cmpxchg() 166 unsigned long prev, tmp, cmp, addr64; ____cmpxchg() local 184 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) ____cmpxchg() 187 return prev; ____cmpxchg() 193 unsigned long prev, cmp; ____cmpxchg() local 207 : "=&r"(prev), "=&r"(cmp), "=m"(*m) ____cmpxchg() 210 return prev; ____cmpxchg() 216 unsigned long prev, cmp; ____cmpxchg() local 230 : "=&r"(prev), "=&r"(cmp), "=m"(*m) ____cmpxchg() 233 return prev; ____cmpxchg()
|
H A D | futex.h | 88 u32 prev; futex_atomic_cmpxchg_inatomic() local 110 : "+r"(ret), "=&r"(prev), "=&r"(cmp) futex_atomic_cmpxchg_inatomic() 114 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
H A D | gct.h | 24 gct6_handle prev; member in struct:__gct6_node
|
/linux-4.4.14/arch/avr32/kernel/ |
H A D | switch_to.S | 15 /* Switch thread context from "prev" to "next", returning "last" 16 * r12 : prev 17 * r11 : &prev->thread + 1
|
/linux-4.4.14/tools/include/linux/ |
H A D | list.h | 19 begin->prev->next = end->next; list_del_range() 20 end->next->prev = begin->prev; list_del_range()
|
/linux-4.4.14/security/selinux/ss/ |
H A D | hashtab.c | 42 struct hashtab_node *prev, *cur, *newnode; hashtab_insert() local 50 prev = NULL; hashtab_insert() 53 prev = cur; hashtab_insert() 65 if (prev) { hashtab_insert() 66 newnode->next = prev->next; hashtab_insert() 67 prev->next = newnode; hashtab_insert()
|
H A D | ebitmap.c | 51 struct ebitmap_node *n, *new, *prev; ebitmap_cpy() local 55 prev = NULL; ebitmap_cpy() 65 if (prev) ebitmap_cpy() 66 prev->next = new; ebitmap_cpy() 69 prev = new; ebitmap_cpy() 258 struct ebitmap_node *n, *prev, *new; ebitmap_set_bit() local 260 prev = NULL; ebitmap_set_bit() 281 if (prev) ebitmap_set_bit() 282 e->highbit = prev->startbit ebitmap_set_bit() 287 if (prev) ebitmap_set_bit() 288 prev->next = n->next; ebitmap_set_bit() 295 prev = n; ebitmap_set_bit() 313 if (prev) { ebitmap_set_bit() 314 new->next = prev->next; ebitmap_set_bit() 315 prev->next = new; ebitmap_set_bit()
|
H A D | mls.c | 37 int i, l, len, head, prev; mls_compute_context_len() local 52 prev = -2; mls_compute_context_len() 55 if (i - prev > 1) { ebitmap_for_each_positive_bit() 57 if (head != prev) { ebitmap_for_each_positive_bit() 58 nm = sym_name(&policydb, SYM_CATS, prev); ebitmap_for_each_positive_bit() 65 prev = i; ebitmap_for_each_positive_bit() 67 if (prev != head) { 68 nm = sym_name(&policydb, SYM_CATS, prev); 92 int i, l, head, prev; mls_sid_to_context() local 111 prev = -2; mls_sid_to_context() 114 if (i - prev > 1) { ebitmap_for_each_positive_bit() 116 if (prev != head) { ebitmap_for_each_positive_bit() 117 if (prev - head > 1) ebitmap_for_each_positive_bit() 121 nm = sym_name(&policydb, SYM_CATS, prev); ebitmap_for_each_positive_bit() 125 if (prev < 0) ebitmap_for_each_positive_bit() 134 prev = i; ebitmap_for_each_positive_bit() 137 if (prev != head) { 138 if (prev - head > 1) 142 nm = sym_name(&policydb, SYM_CATS, prev);
|
H A D | sidtab.c | 36 struct sidtab_node *prev, *cur, *newnode; sidtab_insert() local 44 prev = NULL; sidtab_insert() 47 prev = cur; sidtab_insert() 68 if (prev) { sidtab_insert() 69 newnode->next = prev->next; sidtab_insert() 71 prev->next = newnode; sidtab_insert()
|
/linux-4.4.14/fs/jffs2/ |
H A D | nodelist.c | 28 struct jffs2_full_dirent **prev = list; jffs2_add_fd_to_list() local 32 while ((*prev) && (*prev)->nhash <= new->nhash) { jffs2_add_fd_to_list() 33 if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { jffs2_add_fd_to_list() 35 if (new->version < (*prev)->version) { jffs2_add_fd_to_list() 37 (*prev)->name, (*prev)->ino); jffs2_add_fd_to_list() 42 (*prev)->name, (*prev)->ino); jffs2_add_fd_to_list() 43 new->next = (*prev)->next; jffs2_add_fd_to_list() 46 if ((*prev)->raw) jffs2_add_fd_to_list() 47 jffs2_mark_node_obsolete(c, ((*prev)->raw)); jffs2_add_fd_to_list() 48 jffs2_free_full_dirent(*prev); jffs2_add_fd_to_list() 49 *prev = new; jffs2_add_fd_to_list() 53 prev = &((*prev)->next); jffs2_add_fd_to_list() 55 new->next = *prev; jffs2_add_fd_to_list() 56 *prev = new; jffs2_add_fd_to_list() 386 struct jffs2_node_frag *prev = frag_prev(newfrag); jffs2_add_full_dnode_to_inode() local 390 if (prev->node) jffs2_add_full_dnode_to_inode() 391 mark_ref_normal(prev->node->raw); jffs2_add_full_dnode_to_inode() 438 struct jffs2_inode_cache **prev; jffs2_add_ino_cache() local 446 prev = &c->inocache_list[new->ino % c->inocache_hashsize]; jffs2_add_ino_cache() 448 while ((*prev) && (*prev)->ino < new->ino) { jffs2_add_ino_cache() 449 prev = &(*prev)->next; jffs2_add_ino_cache() 451 new->next = *prev; jffs2_add_ino_cache() 452 *prev = new; jffs2_add_ino_cache() 459 struct jffs2_inode_cache **prev; jffs2_del_ino_cache() local 467 prev = &c->inocache_list[old->ino % c->inocache_hashsize]; jffs2_del_ino_cache() 469 while ((*prev) && (*prev)->ino < old->ino) { jffs2_del_ino_cache() 470 prev = &(*prev)->next; jffs2_del_ino_cache() 472 if ((*prev) == old) { jffs2_del_ino_cache() 473 *prev = old->next; jffs2_del_ino_cache() 529 struct jffs2_node_frag *prev = NULL; jffs2_lookup_node_frag() local 541 if (!prev || frag->ofs > prev->ofs) jffs2_lookup_node_frag() 542 prev = frag; jffs2_lookup_node_frag() 554 if (prev) jffs2_lookup_node_frag() 556 prev->ofs, prev->ofs+prev->size); jffs2_lookup_node_frag() 560 return prev; jffs2_lookup_node_frag()
|
H A D | erase.c | 239 struct jffs2_raw_node_ref **prev; jffs2_remove_node_refs_from_ino_list() local 241 prev = &ref->next_in_ino; jffs2_remove_node_refs_from_ino_list() 245 if (!(*prev)->next_in_ino) { jffs2_remove_node_refs_from_ino_list() 249 ic = (struct jffs2_inode_cache *)(*prev); jffs2_remove_node_refs_from_ino_list() 250 prev = &ic->nodes; jffs2_remove_node_refs_from_ino_list() 254 if (SECTOR_ADDR((*prev)->flash_offset) == jeb->offset) { jffs2_remove_node_refs_from_ino_list() 258 this = *prev; jffs2_remove_node_refs_from_ino_list() 259 *prev = this->next_in_ino; jffs2_remove_node_refs_from_ino_list() 268 prev = &((*prev)->next_in_ino); jffs2_remove_node_refs_from_ino_list()
|
/linux-4.4.14/drivers/video/fbdev/omap2/dss/ |
H A D | dss-of.c | 27 struct device_node *prev) omapdss_of_get_next_port() 34 if (!prev) { omapdss_of_get_next_port() 51 ports = of_get_parent(prev); omapdss_of_get_next_port() 56 port = of_get_next_child(ports, prev); omapdss_of_get_next_port() 61 prev = port; omapdss_of_get_next_port() 73 struct device_node *prev) omapdss_of_get_next_endpoint() 81 ep = of_get_next_child(parent, prev); omapdss_of_get_next_endpoint() 84 prev = ep; omapdss_of_get_next_endpoint() 26 omapdss_of_get_next_port(const struct device_node *parent, struct device_node *prev) omapdss_of_get_next_port() argument 72 omapdss_of_get_next_endpoint(const struct device_node *parent, struct device_node *prev) omapdss_of_get_next_endpoint() argument
|
/linux-4.4.14/arch/nios2/include/asm/ |
H A D | switch_to.h | 17 #define switch_to(prev, next, last) \ 26 : "r" (prev), "r" (next) \
|
H A D | mmu_context.h | 52 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 64 void activate_mm(struct mm_struct *prev, struct mm_struct *next);
|
/linux-4.4.14/fs/btrfs/ |
H A D | extent_map.c | 145 struct rb_node *prev = NULL; __tree_search() local 152 prev = n; __tree_search() 164 orig_prev = prev; __tree_search() 165 while (prev && offset >= extent_map_end(prev_entry)) { __tree_search() 166 prev = rb_next(prev); __tree_search() 167 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search() 169 *prev_ret = prev; __tree_search() 170 prev = orig_prev; __tree_search() 174 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search() 175 while (prev && offset < prev_entry->start) { __tree_search() 176 prev = rb_prev(prev); __tree_search() 177 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search() 179 *next_ret = prev; __tree_search() 185 static int mergable_maps(struct extent_map *prev, struct extent_map *next) mergable_maps() argument 187 if (test_bit(EXTENT_FLAG_PINNED, &prev->flags)) mergable_maps() 194 if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags)) mergable_maps() 197 if (test_bit(EXTENT_FLAG_LOGGING, &prev->flags) || mergable_maps() 206 if (!list_empty(&prev->list) || !list_empty(&next->list)) mergable_maps() 209 if (extent_map_end(prev) == next->start && mergable_maps() 210 prev->flags == next->flags && mergable_maps() 211 prev->bdev == next->bdev && mergable_maps() 213 prev->block_start == EXTENT_MAP_HOLE) || mergable_maps() 215 prev->block_start == EXTENT_MAP_INLINE) || mergable_maps() 217 prev->block_start == EXTENT_MAP_DELALLOC) || mergable_maps() 219 next->block_start == extent_map_block_end(prev)))) { mergable_maps() 365 struct rb_node *prev = NULL; __lookup_extent_mapping() local 369 rb_node = __tree_search(&tree->map, start, &prev, &next); __lookup_extent_mapping() 371 if (prev) __lookup_extent_mapping() 372 rb_node = prev; __lookup_extent_mapping()
|
H A D | ordered-data.c | 81 struct rb_node *prev = NULL; __tree_search() local 88 prev = n; __tree_search() 101 while (prev && file_offset >= entry_end(prev_entry)) { __tree_search() 102 test = rb_next(prev); __tree_search() 110 prev = test; __tree_search() 112 if (prev) __tree_search() 113 prev_entry = rb_entry(prev, struct btrfs_ordered_extent, __tree_search() 115 while (prev && file_offset < entry_end(prev_entry)) { __tree_search() 116 test = rb_prev(prev); __tree_search() 121 prev = test; __tree_search() 123 *prev_ret = prev; __tree_search() 155 struct rb_node *prev = NULL; tree_search() local 165 ret = __tree_search(root, file_offset, &prev); tree_search() 167 ret = prev; tree_search() 441 struct rb_node *prev; btrfs_get_logged_extents() local 445 n = __tree_search(&tree->tree, end, &prev); btrfs_get_logged_extents() 447 n = prev; btrfs_get_logged_extents() 951 struct rb_node *prev = NULL; btrfs_ordered_update_i_size() local 997 prev = tree_search(tree, offset); btrfs_ordered_update_i_size() 1002 if (prev) { btrfs_ordered_update_i_size() 1003 test = rb_entry(prev, struct btrfs_ordered_extent, btrfs_ordered_update_i_size() 1007 node = prev; btrfs_ordered_update_i_size()
|
/linux-4.4.14/include/uapi/linux/netfilter_bridge/ |
H A D | ebt_limit.h | 19 unsigned long prev; member in struct:ebt_limit_info
|
/linux-4.4.14/arch/mips/kernel/ |
H A D | csrc-r4k.c | 47 unsigned int prev, curr, i; rdhwr_count_usable() local 55 for (i = 0, prev = rdhwr_count(); i < 100; i++) { rdhwr_count_usable() 58 if (curr != prev) rdhwr_count_usable() 61 prev = curr; rdhwr_count_usable()
|
/linux-4.4.14/arch/h8300/include/asm/ |
H A D | switch_to.h | 27 * pass prev in a0, next in a1, offset of tss in d1, and whether 34 #define switch_to(prev, next, last) \ 44 : "r" (&(prev->thread)), \ 46 "g" (prev) \
|
/linux-4.4.14/drivers/usb/host/ |
H A D | ohci-mem.c | 105 struct td **prev = &hc->td_hash [TD_HASH_FUNC (td->td_dma)]; td_free() local 107 while (*prev && *prev != td) td_free() 108 prev = &(*prev)->td_hash; td_free() 109 if (*prev) td_free() 110 *prev = td->td_hash; td_free()
|
/linux-4.4.14/arch/m68k/include/asm/ |
H A D | atomic.h | 132 int prev; atomic_cmpxchg() local 135 prev = atomic_read(v); atomic_cmpxchg() 136 if (prev == old) atomic_cmpxchg() 139 return prev; atomic_cmpxchg() 145 int prev; atomic_xchg() local 148 prev = atomic_read(v); atomic_xchg() 151 return prev; atomic_xchg()
|
H A D | switch_to.h | 27 * pass prev in a0, next in a1 30 #define switch_to(prev,next,last) do { \ 31 register void *_prev __asm__ ("a0") = (prev); \
|
/linux-4.4.14/drivers/gpu/drm/nouveau/include/nvif/ |
H A D | list.h | 111 struct list_head *next, *prev; member in struct:list_head 130 list->next = list->prev = list; INIT_LIST_HEAD() 135 struct list_head *prev, struct list_head *next) __list_add() 137 next->prev = entry; __list_add() 139 entry->prev = prev; __list_add() 140 prev->next = entry; __list_add() 182 __list_add(entry, head->prev, head); list_add_tail() 186 __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 188 next->prev = prev; __list_del() 189 prev->next = next; __list_del() 209 __list_del(entry->prev, entry->next); list_del() 215 __list_del(entry->prev, entry->next); list_del_init() 222 __list_del(list->prev, list->next); list_move_tail() 292 list_entry((ptr)->prev, type, member) 334 for (pos = __container_of((head)->prev, pos, member); \ 336 pos = __container_of(pos->member.prev, pos, member)) 344 for (pos = __container_of(pos->member.prev, pos, member); \ 346 pos = __container_of(pos->member.prev, pos, member)) 134 __list_add(struct list_head *entry, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.4.14/arch/tile/include/asm/ |
H A D | switch_to.h | 40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next))) 41 extern struct task_struct *_switch_to(struct task_struct *prev, 45 extern struct task_struct *__switch_to(struct task_struct *prev, 73 struct task_struct *sim_notify_fork(struct task_struct *prev);
|
H A D | mmu_context.h | 91 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 94 if (likely(prev != next)) { switch_mm() 107 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm() 114 check_mm_caching(prev, next); switch_mm()
|
/linux-4.4.14/arch/hexagon/include/asm/ |
H A D | mmu_context.h | 66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 75 if (next->context.generation < prev->context.generation) { switch_mm() 79 next->context.generation = prev->context.generation; switch_mm() 88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 93 switch_mm(prev, next, current_thread_info()->task); activate_mm()
|
H A D | futex.h | 106 int prev; futex_atomic_cmpxchg_inatomic() local 128 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic() 132 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.4.14/drivers/infiniband/hw/mthca/ |
H A D | mthca_mcg.c | 54 * If GID is found in MGM or MGM is empty, *index = *hash, *prev = -1 57 * if GID is found in AMGM, *index = index in AMGM, *prev = index of 60 * If no AMGM exists for given gid, *index = -1, *prev = index of last 65 u16 *hash, int *prev, int *index) find_mgm() 89 *prev = -1; find_mgm() 109 *prev = *index; find_mgm() 126 int index, prev; mthca_multicast_attach() local 138 err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); mthca_multicast_attach() 191 err = mthca_READ_MGM(dev, prev, mailbox); mthca_multicast_attach() 199 err = mthca_WRITE_MGM(dev, prev, mailbox); mthca_multicast_attach() 220 int prev, index; mthca_multicast_detach() local 231 err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); mthca_multicast_detach() 266 if (prev == -1) { mthca_multicast_detach() 291 err = mthca_READ_MGM(dev, prev, mailbox); mthca_multicast_detach() 299 err = mthca_WRITE_MGM(dev, prev, mailbox); mthca_multicast_detach() 63 find_mgm(struct mthca_dev *dev, u8 *gid, struct mthca_mailbox *mgm_mailbox, u16 *hash, int *prev, int *index) find_mgm() argument
|
/linux-4.4.14/tools/perf/scripts/python/ |
H A D | sched-migration.py | 102 def sched_switch(self, prev, prev_state, next): 106 and prev in self.tasks: 110 event = RunqueueEventSleep(prev) 113 if prev in self.tasks: 115 next_tasks.remove(prev) 117 next_tasks.append(prev) 160 def __init__(self, start, prev): 162 self.prev = prev 166 if prev is not None: 167 self.total_load = prev.total_load 168 self.rqs = prev.rqs.copy() 177 def sched_switch(self, ts_list, prev, prev_state, next, cpu): 178 old_rq = self.prev.rqs[cpu] 179 new_rq = old_rq.sched_switch(prev, prev_state, next) 192 old_rq = self.prev.rqs[old_cpu] 197 new_rq = self.prev.rqs[new_cpu] 209 old_rq = self.prev.rqs[cpu] 339 print "Sched switch event rejected ts: %s cpu: %d prev: %s(%d) next: %s(%d)" % \
|
/linux-4.4.14/drivers/gpu/drm/nouveau/ |
H A D | nv17_fence.c | 34 struct nouveau_channel *prev, struct nouveau_channel *chan) nv17_fence_sync() 36 struct nouveau_cli *cli = (void *)prev->user.client; nv17_fence_sync() 50 ret = RING_SPACE(prev, 5); nv17_fence_sync() 52 BEGIN_NV04(prev, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 4); nv17_fence_sync() 53 OUT_RING (prev, fctx->sema.handle); nv17_fence_sync() 54 OUT_RING (prev, 0); nv17_fence_sync() 55 OUT_RING (prev, value + 0); nv17_fence_sync() 56 OUT_RING (prev, value + 1); nv17_fence_sync() 57 FIRE_RING (prev); nv17_fence_sync() 33 nv17_fence_sync(struct nouveau_fence *fence, struct nouveau_channel *prev, struct nouveau_channel *chan) nv17_fence_sync() argument
|
H A D | nouveau_bo.h | 53 struct nouveau_bo *prev; nouveau_bo_ref() local 57 prev = *pnvbo; nouveau_bo_ref() 60 if (prev) { nouveau_bo_ref() 61 struct ttm_buffer_object *bo = &prev->bo; nouveau_bo_ref()
|
H A D | nouveau_fence.c | 410 struct nouveau_channel *prev = NULL; nouveau_fence_sync() local 416 prev = rcu_dereference(f->channel); nouveau_fence_sync() 417 if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) nouveau_fence_sync() 432 struct nouveau_channel *prev = NULL; nouveau_fence_sync() local 441 prev = rcu_dereference(f->channel); nouveau_fence_sync() 442 if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) nouveau_fence_sync()
|
/linux-4.4.14/drivers/char/agp/ |
H A D | frontend.c | 68 struct agp_memory *prev; agp_remove_from_pool() local 76 prev = temp->prev; agp_remove_from_pool() 78 if (prev != NULL) { agp_remove_from_pool() 79 prev->next = next; agp_remove_from_pool() 81 next->prev = prev; agp_remove_from_pool() 86 next->prev = NULL; agp_remove_from_pool() 201 struct agp_memory *prev; agp_insert_into_pool() local 203 prev = agp_fe.current_controller->pool; agp_insert_into_pool() 205 if (prev != NULL) { agp_insert_into_pool() 206 prev->prev = temp; agp_insert_into_pool() 207 temp->next = prev; agp_insert_into_pool() 232 struct agp_file_private *prev; agp_insert_file_private() local 234 prev = agp_fe.file_priv_list; agp_insert_file_private() 236 if (prev != NULL) agp_insert_file_private() 237 prev->prev = priv; agp_insert_file_private() 238 priv->next = prev; agp_insert_file_private() 245 struct agp_file_private *prev; agp_remove_file_private() local 248 prev = priv->prev; agp_remove_file_private() 250 if (prev != NULL) { agp_remove_file_private() 251 prev->next = next; agp_remove_file_private() 254 next->prev = prev; agp_remove_file_private() 258 next->prev = NULL; agp_remove_file_private() 328 prev_controller->prev = controller; agp_insert_controller() 377 prev_controller = controller->prev; agp_remove_controller() 383 next_controller->prev = prev_controller; agp_remove_controller() 387 next_controller->prev = NULL; agp_remove_controller() 507 prev_client->prev = client; agp_insert_client() 541 prev_client = client->prev; agp_remove_client() 547 next_client->prev = prev_client; agp_remove_client() 551 next_client->prev = NULL; agp_remove_client()
|
/linux-4.4.14/crypto/ |
H A D | xcbc.c | 53 * | prev (block size) 87 u8 *prev = PTR_ALIGN(&ctx->ctx[0], alignmask + 1) + bs; crypto_xcbc_digest_init() local 90 memset(prev, 0, bs); crypto_xcbc_digest_init() 105 u8 *prev = odds + bs; crypto_xcbc_digest_update() local 119 crypto_xor(prev, odds, bs); crypto_xcbc_digest_update() 120 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_xcbc_digest_update() 127 crypto_xor(prev, p, bs); crypto_xcbc_digest_update() 128 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_xcbc_digest_update() 152 u8 *prev = odds + bs; crypto_xcbc_digest_final() local 169 crypto_xor(prev, odds, bs); crypto_xcbc_digest_final() 170 crypto_xor(prev, consts + offset, bs); crypto_xcbc_digest_final() 172 crypto_cipher_encrypt_one(tfm, out, prev); crypto_xcbc_digest_final()
|
H A D | cmac.c | 46 * | prev (block size) 113 u8 *prev = PTR_ALIGN((void *)ctx->ctx, alignmask + 1) + bs; crypto_cmac_digest_init() local 116 memset(prev, 0, bs); crypto_cmac_digest_init() 131 u8 *prev = odds + bs; crypto_cmac_digest_update() local 145 crypto_xor(prev, odds, bs); crypto_cmac_digest_update() 146 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_cmac_digest_update() 153 crypto_xor(prev, p, bs); crypto_cmac_digest_update() 154 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_cmac_digest_update() 178 u8 *prev = odds + bs; crypto_cmac_digest_final() local 195 crypto_xor(prev, odds, bs); crypto_cmac_digest_final() 196 crypto_xor(prev, consts + offset, bs); crypto_cmac_digest_final() 198 crypto_cipher_encrypt_one(tfm, out, prev); crypto_cmac_digest_final()
|
/linux-4.4.14/sound/isa/gus/ |
H A D | gus_mem.c | 54 nblock->prev = pblock->prev; snd_gf1_mem_xalloc() 56 pblock->prev = nblock; snd_gf1_mem_xalloc() 60 nblock->prev->next = nblock; snd_gf1_mem_xalloc() 68 nblock->prev = NULL; snd_gf1_mem_xalloc() 71 nblock->prev = alloc->last; snd_gf1_mem_xalloc() 88 block->next->prev = NULL; snd_gf1_mem_xfree() 90 block->prev->next = block->next; snd_gf1_mem_xfree() 92 block->next->prev = block->prev; snd_gf1_mem_xfree() 95 alloc->last = block->prev; snd_gf1_mem_xfree() 96 if (block->prev) snd_gf1_mem_xfree() 97 block->prev->next = NULL; snd_gf1_mem_xfree() 99 block->next->prev = block->prev; snd_gf1_mem_xfree() 100 if (block->prev) snd_gf1_mem_xfree() 101 block->prev->next = block->next; snd_gf1_mem_xfree() 154 block->prev = block->next = NULL; snd_gf1_mem_find()
|
/linux-4.4.14/arch/sparc/lib/ |
H A D | atomic32.c | 159 u32 prev; __cmpxchg_u32() local 162 if ((prev = *ptr) == old) __cmpxchg_u32() 166 return (unsigned long)prev; __cmpxchg_u32() 173 u32 prev; __xchg_u32() local 176 prev = *ptr; __xchg_u32() 180 return (unsigned long)prev; __xchg_u32()
|
H A D | memcpy.S | 100 #define SMOVE_CHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ 106 or %t5, %prev, %t5; \ 107 sll %t1, shil, %prev; \ 112 or %t1, %prev, %t1; \ 115 sll %t3, shil, %prev; \ 118 #define SMOVE_ALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ 124 or %t4, %prev, %t0; \ 125 sll %t1, shil, %prev; \ 130 or %t4, %prev, %t2; \ 131 sll %t3, shil, %prev; \
|
/linux-4.4.14/arch/powerpc/platforms/cell/spufs/ |
H A D | switch.c | 1791 static int quiece_spu(struct spu_state *prev, struct spu *spu) quiece_spu() argument 1803 if (check_spu_isolate(prev, spu)) { /* Step 2. */ quiece_spu() 1806 disable_interrupts(prev, spu); /* Step 3. */ quiece_spu() 1807 set_watchdog_timer(prev, spu); /* Step 4. */ quiece_spu() 1808 inhibit_user_access(prev, spu); /* Step 5. */ quiece_spu() 1809 if (check_spu_isolate(prev, spu)) { /* Step 6. */ quiece_spu() 1812 set_switch_pending(prev, spu); /* Step 7. */ quiece_spu() 1813 save_mfc_cntl(prev, spu); /* Step 8. */ quiece_spu() 1814 save_spu_runcntl(prev, spu); /* Step 9. */ quiece_spu() 1815 save_mfc_sr1(prev, spu); /* Step 10. */ quiece_spu() 1816 save_spu_status(prev, spu); /* Step 11. */ quiece_spu() 1817 save_mfc_stopped_status(prev, spu); /* Step 12. */ quiece_spu() 1818 halt_mfc_decr(prev, spu); /* Step 13. */ quiece_spu() 1819 save_timebase(prev, spu); /* Step 14. */ quiece_spu() 1820 remove_other_spu_access(prev, spu); /* Step 15. */ quiece_spu() 1821 do_mfc_mssync(prev, spu); /* Step 16. */ quiece_spu() 1822 issue_mfc_tlbie(prev, spu); /* Step 17. */ quiece_spu() 1823 handle_pending_interrupts(prev, spu); /* Step 18. */ quiece_spu() 1828 static void save_csa(struct spu_state *prev, struct spu *spu) save_csa() argument 1835 save_mfc_queues(prev, spu); /* Step 19. */ save_csa() 1836 save_ppu_querymask(prev, spu); /* Step 20. */ save_csa() 1837 save_ppu_querytype(prev, spu); /* Step 21. */ save_csa() 1838 save_ppu_tagstatus(prev, spu); /* NEW. */ save_csa() 1839 save_mfc_csr_tsq(prev, spu); /* Step 22. */ save_csa() 1840 save_mfc_csr_cmd(prev, spu); /* Step 23. */ save_csa() 1841 save_mfc_csr_ato(prev, spu); /* Step 24. */ save_csa() 1842 save_mfc_tclass_id(prev, spu); /* Step 25. */ save_csa() 1843 set_mfc_tclass_id(prev, spu); /* Step 26. */ save_csa() 1844 save_mfc_cmd(prev, spu); /* Step 26a - moved from 44. */ save_csa() 1845 purge_mfc_queue(prev, spu); /* Step 27. */ save_csa() 1846 wait_purge_complete(prev, spu); /* Step 28. */ save_csa() 1847 setup_mfc_sr1(prev, spu); /* Step 30. */ save_csa() 1848 save_spu_npc(prev, spu); /* Step 31. */ save_csa() 1849 save_spu_privcntl(prev, spu); /* Step 32. */ save_csa() 1850 reset_spu_privcntl(prev, spu); /* Step 33. */ save_csa() 1851 save_spu_lslr(prev, spu); /* Step 34. */ save_csa() 1852 reset_spu_lslr(prev, spu); /* Step 35. */ save_csa() 1853 save_spu_cfg(prev, spu); /* Step 36. */ save_csa() 1854 save_pm_trace(prev, spu); /* Step 37. */ save_csa() 1855 save_mfc_rag(prev, spu); /* Step 38. */ save_csa() 1856 save_ppu_mb_stat(prev, spu); /* Step 39. */ save_csa() 1857 save_ppu_mb(prev, spu); /* Step 40. */ save_csa() 1858 save_ppuint_mb(prev, spu); /* Step 41. */ save_csa() 1859 save_ch_part1(prev, spu); /* Step 42. */ save_csa() 1860 save_spu_mb(prev, spu); /* Step 43. */ save_csa() 1861 reset_ch(prev, spu); /* Step 45. */ save_csa() 1864 static void save_lscsa(struct spu_state *prev, struct spu *spu) save_lscsa() argument 1872 resume_mfc_queue(prev, spu); /* Step 46. */ save_lscsa() 1874 setup_mfc_slbs(prev, spu, spu_save_code, sizeof(spu_save_code)); save_lscsa() 1875 set_switch_active(prev, spu); /* Step 48. */ save_lscsa() 1876 enable_interrupts(prev, spu); /* Step 49. */ save_lscsa() 1877 save_ls_16kb(prev, spu); /* Step 50. */ save_lscsa() 1878 set_spu_npc(prev, spu); /* Step 51. */ save_lscsa() 1879 set_signot1(prev, spu); /* Step 52. */ save_lscsa() 1880 set_signot2(prev, spu); /* Step 53. */ save_lscsa() 1881 send_save_code(prev, spu); /* Step 54. */ save_lscsa() 1882 set_ppu_querymask(prev, spu); /* Step 55. */ save_lscsa() 1883 wait_tag_complete(prev, spu); /* Step 56. */ save_lscsa() 1884 wait_spu_stopped(prev, spu); /* Step 57. */ save_lscsa() 1932 static void harvest(struct spu_state *prev, struct spu *spu) harvest() argument 1940 disable_interrupts(prev, spu); /* Step 2. */ harvest() 1941 inhibit_user_access(prev, spu); /* Step 3. */ harvest() 1942 terminate_spu_app(prev, spu); /* Step 4. */ harvest() 1943 set_switch_pending(prev, spu); /* Step 5. */ harvest() 1945 remove_other_spu_access(prev, spu); /* Step 6. */ harvest() 1946 suspend_mfc_and_halt_decr(prev, spu); /* Step 7. */ harvest() 1947 wait_suspend_mfc_complete(prev, spu); /* Step 8. */ harvest() 1948 if (!suspend_spe(prev, spu)) /* Step 9. */ harvest() 1949 clear_spu_status(prev, spu); /* Step 10. */ harvest() 1950 do_mfc_mssync(prev, spu); /* Step 11. */ harvest() 1951 issue_mfc_tlbie(prev, spu); /* Step 12. */ harvest() 1952 handle_pending_interrupts(prev, spu); /* Step 13. */ harvest() 1953 purge_mfc_queue(prev, spu); /* Step 14. */ harvest() 1954 wait_purge_complete(prev, spu); /* Step 15. */ harvest() 1955 reset_spu_privcntl(prev, spu); /* Step 16. */ harvest() 1956 reset_spu_lslr(prev, spu); /* Step 17. */ harvest() 1957 setup_mfc_sr1(prev, spu); /* Step 18. */ harvest() 1959 reset_ch_part1(prev, spu); /* Step 20. */ harvest() 1960 reset_ch_part2(prev, spu); /* Step 21. */ harvest() 1961 enable_interrupts(prev, spu); /* Step 22. */ harvest() 1962 set_switch_active(prev, spu); /* Step 23. */ harvest() 1963 set_mfc_tclass_id(prev, spu); /* Step 24. */ harvest() 1964 resume_mfc_queue(prev, spu); /* Step 25. */ harvest() 2038 static int __do_spu_save(struct spu_state *prev, struct spu *spu) __do_spu_save() argument 2054 rc = quiece_spu(prev, spu); /* Steps 2-16. */ __do_spu_save() 2059 harvest(prev, spu); __do_spu_save() 2065 save_csa(prev, spu); /* Steps 17-43. */ __do_spu_save() 2066 save_lscsa(prev, spu); /* Steps 44-53. */ __do_spu_save() 2067 return check_save_status(prev, spu); /* Step 54. */ __do_spu_save() 2103 * @prev: pointer to SPU context save area, to be saved. 2108 int spu_save(struct spu_state *prev, struct spu *spu) spu_save() argument 2113 rc = __do_spu_save(prev, spu); /* Steps 2-53. */ spu_save()
|
/linux-4.4.14/arch/s390/lib/ |
H A D | spinlock.c | 186 void _raw_write_lock_wait(arch_rwlock_t *rw, unsigned int prev) _raw_write_lock_wait() argument 202 prev = __RAW_LOCK(&rw->lock, 0x80000000, __RAW_OP_OR); _raw_write_lock_wait() 203 old = prev; _raw_write_lock_wait() 205 if ((old & 0x7fffffff) == 0 && (int) prev >= 0) _raw_write_lock_wait() 217 unsigned int owner, old, prev; _raw_write_lock_wait() local 220 prev = 0x80000000; _raw_write_lock_wait() 232 prev = old; _raw_write_lock_wait() 235 if ((old & 0x7fffffff) == 0 && (int) prev >= 0) _raw_write_lock_wait()
|
/linux-4.4.14/tools/perf/ |
H A D | perf-completion.sh | 73 prev) 74 prev=${words_[$cword_-1]} 162 elif [[ $prev == @("-e"|"--event") && 232 local _ret=1 cur cword prev 234 prev=${words[CURRENT-1]} 248 local cur words cword prev 250 _get_comp_words_by_ref -n =: cur words cword prev 252 __perf_get_comp_words_by_ref -n =: cur words cword prev
|
H A D | builtin-timechart.c | 726 struct io_sample *prev; pid_begin_io_sample() local 737 prev = c->io_samples; pid_begin_io_sample() 739 if (prev && prev->start_time && !prev->end_time) { pid_begin_io_sample() 745 c->io_samples = prev->next; pid_begin_io_sample() 746 free(prev); pid_begin_io_sample() 770 struct io_sample *sample, *prev; pid_end_io_sample() local 794 prev = sample->next; pid_end_io_sample() 800 if (prev && sample->start_time < prev->end_time) { pid_end_io_sample() 801 if (prev->err) /* try to make errors more visible */ pid_end_io_sample() 802 sample->start_time = prev->end_time; pid_end_io_sample() 804 prev->end_time = sample->start_time; pid_end_io_sample() 821 if (prev && pid_end_io_sample() 822 prev->type == sample->type && pid_end_io_sample() 823 prev->err == sample->err && pid_end_io_sample() 824 prev->fd == sample->fd && pid_end_io_sample() 825 prev->end_time + tchart->merge_dist >= sample->start_time) { pid_end_io_sample() 827 sample->bytes += prev->bytes; pid_end_io_sample() 828 sample->merges += prev->merges + 1; pid_end_io_sample() 830 sample->start_time = prev->start_time; pid_end_io_sample() 831 sample->next = prev->next; pid_end_io_sample() 832 free(prev); pid_end_io_sample() 968 struct per_pid *new_list, *p, *cursor, *prev; sort_pids() local 983 prev = NULL; sort_pids() 989 if (prev) { sort_pids() 990 p->next = prev->next; sort_pids() 991 prev->next = p; sort_pids() 1002 prev = cursor; sort_pids() 1005 prev->next = p; sort_pids()
|
/linux-4.4.14/include/uapi/linux/netfilter/ |
H A D | xt_limit.h | 18 unsigned long prev; /* moved to xt_limit_priv */ member in struct:xt_rateinfo
|
/linux-4.4.14/arch/um/include/asm/ |
H A D | mmu_context.h | 50 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 55 if(prev != next){ switch_mm() 56 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
|
/linux-4.4.14/kernel/sched/ |
H A D | idle_task.c | 27 pick_next_task_idle(struct rq *rq, struct task_struct *prev) pick_next_task_idle() argument 29 put_prev_task(rq, prev); pick_next_task_idle() 48 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) put_prev_task_idle() argument
|
H A D | stop_task.c | 27 pick_next_task_stop(struct rq *rq, struct task_struct *prev) pick_next_task_stop() argument 34 put_prev_task(rq, prev); pick_next_task_stop() 58 static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) put_prev_task_stop() argument
|
H A D | cputime.c | 392 void vtime_common_task_switch(struct task_struct *prev) vtime_common_task_switch() argument 394 if (is_idle_task(prev)) vtime_common_task_switch() 395 vtime_account_idle(prev); vtime_common_task_switch() 397 vtime_account_system(prev); vtime_common_task_switch() 400 vtime_account_user(prev); vtime_common_task_switch() 402 arch_vtime_task_switch(prev); vtime_common_task_switch() 579 struct prev_cputime *prev, cputime_adjust() 586 raw_spin_lock_irqsave(&prev->lock, flags); cputime_adjust() 597 if (prev->stime + prev->utime >= rtime) cputime_adjust() 625 if (stime < prev->stime) cputime_adjust() 626 stime = prev->stime; cputime_adjust() 633 if (utime < prev->utime) { cputime_adjust() 634 utime = prev->utime; cputime_adjust() 639 prev->stime = stime; cputime_adjust() 640 prev->utime = utime; cputime_adjust() 642 *ut = prev->utime; cputime_adjust() 643 *st = prev->stime; cputime_adjust() 644 raw_spin_unlock_irqrestore(&prev->lock, flags); cputime_adjust() 764 void arch_vtime_task_switch(struct task_struct *prev) arch_vtime_task_switch() argument 766 write_seqlock(&prev->vtime_seqlock); arch_vtime_task_switch() 767 prev->vtime_snap_whence = VTIME_SLEEPING; arch_vtime_task_switch() 768 write_sequnlock(&prev->vtime_seqlock); arch_vtime_task_switch() 578 cputime_adjust(struct task_cputime *curr, struct prev_cputime *prev, cputime_t *ut, cputime_t *st) cputime_adjust() argument
|
H A D | stats.h | 128 * the idle task.) We are only called when prev != next. 132 struct task_struct *prev, struct task_struct *next) __sched_info_switch() 135 * prev now departs the cpu. It's not interesting to record __sched_info_switch() 139 if (prev != rq->idle) __sched_info_switch() 140 sched_info_depart(rq, prev); __sched_info_switch() 147 struct task_struct *prev, struct task_struct *next) sched_info_switch() 150 __sched_info_switch(rq, prev, next); sched_info_switch() 131 __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) __sched_info_switch() argument 146 sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) sched_info_switch() argument
|
H A D | auto_group.c | 128 struct autogroup *prev; autogroup_move_group() local 134 prev = p->signal->autogroup; autogroup_move_group() 135 if (prev == ag) { autogroup_move_group() 149 autogroup_kref_put(prev); autogroup_move_group()
|
/linux-4.4.14/arch/openrisc/include/asm/ |
H A D | mmu_context.h | 26 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 31 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.4.14/drivers/isdn/divert/ |
H A D | isdn_divert.c | 30 struct call_struc *prev; member in struct:call_struc 38 struct deflect_struc *next, *prev; member in struct:deflect_struc 93 if (cs->prev) deflect_timer_expire() 94 cs->prev->next = cs->next; /* forward link */ deflect_timer_expire() 98 cs->next->prev = cs->prev; /* back link */ deflect_timer_expire() 191 cs->prev = NULL; cf_command() 295 ds->prev = table_tail; /* previous entry */ insertrule() 297 if (ds->prev) insertrule() 298 ds->prev->next = ds; /* last forward */ insertrule() 304 ds->prev = ds1->prev; /* prev entry */ insertrule() 305 ds1->prev = ds; /* backward chain old element */ insertrule() 306 if (!ds->prev) insertrule() 350 ds->next->prev = ds->prev; /* backward chain */ deleterule() 352 table_tail = ds->prev; /* end of chain */ deleterule() 354 if (ds->prev) deleterule() 355 ds->prev->next = ds->next; /* forward chain */ deleterule() 474 cs->prev = NULL; isdn_divert_icall() 518 cs->prev = NULL; isdn_divert_icall() 756 if (cs1->prev) prot_stat_callback() 757 cs1->prev->next = cs1->next; /* forward link */ prot_stat_callback() 761 cs1->next->prev = cs1->prev; /* back link */ prot_stat_callback() 812 if (cs1->prev) isdn_divert_stat_callback() 813 cs1->prev->next = cs1->next; /* forward link */ isdn_divert_stat_callback() 817 cs1->next->prev = cs1->prev; /* back link */ isdn_divert_stat_callback()
|
/linux-4.4.14/net/ceph/ |
H A D | pagelist.c | 11 struct page *page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_unmap_tail() 119 c->page_lru = pl->head.prev; ceph_pagelist_set_cursor() 137 while (pl->head.prev != c->page_lru) { ceph_pagelist_truncate() 138 page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_truncate() 145 page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_truncate()
|
/linux-4.4.14/net/netfilter/ |
H A D | xt_limit.c | 21 unsigned long prev; member in struct:xt_limit_priv 43 `prev' tracks the last packet hit: you gain one credit per jiffy. 76 priv->credit += (now - xchg(&priv->prev, now)) * CREDITS_PER_JIFFY; limit_mt() 123 priv->prev = jiffies; limit_mt_check() 144 compat_ulong_t prev; member in struct:compat_xt_rateinfo 151 /* To keep the full "prev" timestamp, the upper 32 bits are stored in the 159 .prev = cm->prev | (unsigned long)cm->master << 32, limit_mt_compat_from_user() 173 .prev = m->prev, limit_mt_compat_to_user() 177 .master = m->prev >> 32, limit_mt_compat_to_user()
|
/linux-4.4.14/net/netfilter/ipset/ |
H A D | ip_set_list_set.c | 197 struct set_elem *e, *next, *prev = NULL; list_set_utest() local 205 prev = e; list_set_utest() 216 ret = prev && prev->id == d->refid; list_set_utest() 244 struct set_elem *e, *n, *prev, *next; list_set_uadd() local 251 n = prev = next = NULL; list_set_uadd() 263 prev = e; list_set_uadd() 268 (d->before < 0 && !prev)) list_set_uadd() 290 /* Insert before prev element */ list_set_uadd() 291 if (prev->list.prev != &map->members) list_set_uadd() 292 n = list_prev_entry(prev, list); list_set_uadd() 310 else if (prev) list_set_uadd() 311 list_add_rcu(&e->list, &prev->list); list_set_uadd() 324 struct set_elem *e, *next, *prev = NULL; list_set_udel() local 331 prev = e; list_set_udel() 341 if (!prev || prev->id != d->refid) list_set_udel()
|
/linux-4.4.14/arch/powerpc/boot/ |
H A D | ops.h | 47 void *(*find_node_by_prop_value)(const void *prev, 50 void *(*find_node_by_compatible)(const void *prev, 146 static inline void *find_node_by_prop_value(const void *prev, find_node_by_prop_value() argument 151 return dt_ops.find_node_by_prop_value(prev, propname, find_node_by_prop_value() 157 static inline void *find_node_by_prop_value_str(const void *prev, find_node_by_prop_value_str() argument 161 return find_node_by_prop_value(prev, propname, propval, find_node_by_prop_value_str() 165 static inline void *find_node_by_devtype(const void *prev, find_node_by_devtype() argument 168 return find_node_by_prop_value_str(prev, "device_type", type); find_node_by_devtype() 184 static inline void *find_node_by_compatible(const void *prev, find_node_by_compatible() argument 188 return dt_ops.find_node_by_compatible(prev, compat); find_node_by_compatible()
|
H A D | libfdt-wrapper.c | 131 static void *fdt_wrapper_find_node_by_prop_value(const void *prev, fdt_wrapper_find_node_by_prop_value() argument 136 int offset = fdt_node_offset_by_prop_value(fdt, devp_offset_find(prev), fdt_wrapper_find_node_by_prop_value() 141 static void *fdt_wrapper_find_node_by_compatible(const void *prev, fdt_wrapper_find_node_by_compatible() argument 144 int offset = fdt_node_offset_by_compatible(fdt, devp_offset_find(prev), fdt_wrapper_find_node_by_compatible()
|
/linux-4.4.14/arch/x86/kernel/cpu/ |
H A D | perf_event_msr.c | 154 u64 prev, now; msr_event_update() local 159 prev = local64_read(&event->hw.prev_count); msr_event_update() 162 if (local64_cmpxchg(&event->hw.prev_count, prev, now) != prev) msr_event_update() 165 delta = now - prev; msr_event_update() 169 local64_add(now - prev, &event->count); msr_event_update()
|
/linux-4.4.14/sound/synth/ |
H A D | util_mem.c | 101 return __snd_util_memblk_new(hdr, units, p->prev); __snd_util_mem_alloc() 107 * the block is linked next to prev 111 struct list_head *prev) __snd_util_memblk_new() 120 if (prev == &hdr->block) __snd_util_memblk_new() 123 struct snd_util_memblk *p = get_memblk(prev); __snd_util_memblk_new() 127 list_add(&blk->list, prev); __snd_util_memblk_new() 110 __snd_util_memblk_new(struct snd_util_memhdr *hdr, unsigned int units, struct list_head *prev) __snd_util_memblk_new() argument
|
/linux-4.4.14/drivers/gpu/drm/ |
H A D | drm_lock.c | 180 unsigned int old, new, prev; drm_lock_take() local 193 prev = cmpxchg(lock, old, new); drm_lock_take() 194 } while (prev != old); drm_lock_take() 229 unsigned int old, new, prev; drm_lock_transfer() local 236 prev = cmpxchg(lock, old, new); drm_lock_transfer() 237 } while (prev != old); drm_lock_transfer() 254 unsigned int old, new, prev; drm_legacy_lock_free() local 269 prev = cmpxchg(lock, old, new); drm_legacy_lock_free() 270 } while (prev != old); drm_legacy_lock_free() 315 unsigned int old, prev; drm_legacy_idlelock_release() local 323 prev = cmpxchg(lock, old, DRM_KERNEL_CONTEXT); drm_legacy_idlelock_release() 324 } while (prev != old); drm_legacy_idlelock_release()
|
/linux-4.4.14/net/bridge/netfilter/ |
H A D | ebt_limit.c | 41 info->credit += (now - xchg(&info->prev, now)) * CREDITS_PER_JIFFY; ebt_limit_mt() 81 info->prev = jiffies; ebt_limit_mt_check() 96 compat_ulong_t prev; member in struct:ebt_compat_limit_info
|
/linux-4.4.14/arch/metag/kernel/ |
H A D | process.c | 301 struct task_struct *__sched __switch_to(struct task_struct *prev, __switch_to() argument 307 to.Switch.pPara = prev; __switch_to() 310 if (prev->thread.user_flags & TBICTX_FPAC_BIT) { __switch_to() 311 struct pt_regs *regs = task_pt_regs(prev); __switch_to() 314 state.Sig.SaveMask = prev->thread.user_flags; __switch_to() 317 if (!prev->thread.fpu_context) __switch_to() 318 alloc_fpu_context(&prev->thread); __switch_to() 319 if (prev->thread.fpu_context) __switch_to() 320 __TBICtxFPUSave(state, prev->thread.fpu_context); __switch_to() 326 if (prev->thread.fpu_context) __switch_to() 327 prev->thread.fpu_context->needs_restore = true; __switch_to() 331 from = __TBISwitch(to, &prev->thread.kernel_context); __switch_to()
|
/linux-4.4.14/drivers/usb/image/ |
H A D | microtek.h | 31 struct mts_desc *prev; member in struct:mts_desc
|
/linux-4.4.14/drivers/media/pci/zoran/ |
H A D | videocodec.c | 172 struct attached_list *a, *prev; videocodec_detach() local 192 prev = NULL; videocodec_detach() 208 if (prev == NULL) { videocodec_detach() 213 prev->next = a->next; videocodec_detach() 223 prev = a; videocodec_detach() 271 struct codec_list *prev = NULL, *h = codeclist_top; videocodec_unregister() local 300 if (prev == NULL) { videocodec_unregister() 305 prev->next = h->next; videocodec_unregister() 312 prev = h; videocodec_unregister()
|
/linux-4.4.14/drivers/input/joystick/ |
H A D | joydump.c | 52 struct joydump *dump, *prev; /* one entry each */ joydump_connect() local 123 prev = dump; joydump_connect() 132 for (i = 1; i < t; i++, dump++, prev++) { joydump_connect() 134 i, dump->time - prev->time); joydump_connect()
|
/linux-4.4.14/drivers/net/ethernet/sfc/ |
H A D | mdio_10g.c | 235 struct ethtool_cmd prev = { .cmd = ETHTOOL_GSET }; efx_mdio_set_settings() local 237 efx->phy_op->get_settings(efx, &prev); efx_mdio_set_settings() 239 if (ecmd->advertising == prev.advertising && efx_mdio_set_settings() 240 ethtool_cmd_speed(ecmd) == ethtool_cmd_speed(&prev) && efx_mdio_set_settings() 241 ecmd->duplex == prev.duplex && efx_mdio_set_settings() 242 ecmd->port == prev.port && efx_mdio_set_settings() 243 ecmd->autoneg == prev.autoneg) efx_mdio_set_settings() 247 if (prev.port != PORT_TP || ecmd->port != PORT_TP) efx_mdio_set_settings() 252 (ecmd->advertising | SUPPORTED_Autoneg) & ~prev.supported) efx_mdio_set_settings()
|
/linux-4.4.14/arch/arm/mach-omap2/ |
H A D | sdrc2xxx.c | 85 u32 prev = curr_perf_level; omap2xxx_sdrc_reprogram() local 89 return prev; omap2xxx_sdrc_reprogram() 96 return prev; omap2xxx_sdrc_reprogram() 113 return prev; omap2xxx_sdrc_reprogram()
|
/linux-4.4.14/drivers/acpi/ |
H A D | nfit.c | 233 struct nfit_table_prev *prev, add_spa() 240 list_for_each_entry(nfit_spa, &prev->spas, list) { add_spa() 260 struct nfit_table_prev *prev, add_memdev() 267 list_for_each_entry(nfit_memdev, &prev->memdevs, list) add_memdev() 286 struct nfit_table_prev *prev, add_dcr() 293 list_for_each_entry(nfit_dcr, &prev->dcrs, list) add_dcr() 311 struct nfit_table_prev *prev, add_bdw() 318 list_for_each_entry(nfit_bdw, &prev->bdws, list) add_bdw() 336 struct nfit_table_prev *prev, add_idt() 343 list_for_each_entry(nfit_idt, &prev->idts, list) add_idt() 361 struct nfit_table_prev *prev, add_flush() 368 list_for_each_entry(nfit_flush, &prev->flushes, list) add_flush() 386 struct nfit_table_prev *prev, void *table, const void *end) add_table() 404 if (!add_spa(acpi_desc, prev, table)) add_table() 408 if (!add_memdev(acpi_desc, prev, table)) add_table() 412 if (!add_dcr(acpi_desc, prev, table)) add_table() 416 if (!add_bdw(acpi_desc, prev, table)) add_table() 420 if (!add_idt(acpi_desc, prev, table)) add_table() 424 if (!add_flush(acpi_desc, prev, table)) add_table() 1612 struct nfit_table_prev *prev) acpi_nfit_check_deletions() 1616 if (!list_empty(&prev->spas) || acpi_nfit_check_deletions() 1617 !list_empty(&prev->memdevs) || acpi_nfit_check_deletions() 1618 !list_empty(&prev->dcrs) || acpi_nfit_check_deletions() 1619 !list_empty(&prev->bdws) || acpi_nfit_check_deletions() 1620 !list_empty(&prev->idts) || acpi_nfit_check_deletions() 1621 !list_empty(&prev->flushes)) { acpi_nfit_check_deletions() 1631 struct nfit_table_prev prev; acpi_nfit_init() local 1638 INIT_LIST_HEAD(&prev.spas); acpi_nfit_init() 1639 INIT_LIST_HEAD(&prev.memdevs); acpi_nfit_init() 1640 INIT_LIST_HEAD(&prev.dcrs); acpi_nfit_init() 1641 INIT_LIST_HEAD(&prev.bdws); acpi_nfit_init() 1642 INIT_LIST_HEAD(&prev.idts); acpi_nfit_init() 1643 INIT_LIST_HEAD(&prev.flushes); acpi_nfit_init() 1645 list_cut_position(&prev.spas, &acpi_desc->spas, acpi_nfit_init() 1646 acpi_desc->spas.prev); acpi_nfit_init() 1647 list_cut_position(&prev.memdevs, &acpi_desc->memdevs, acpi_nfit_init() 1648 acpi_desc->memdevs.prev); acpi_nfit_init() 1649 list_cut_position(&prev.dcrs, &acpi_desc->dcrs, acpi_nfit_init() 1650 acpi_desc->dcrs.prev); acpi_nfit_init() 1651 list_cut_position(&prev.bdws, &acpi_desc->bdws, acpi_nfit_init() 1652 acpi_desc->bdws.prev); acpi_nfit_init() 1653 list_cut_position(&prev.idts, &acpi_desc->idts, acpi_nfit_init() 1654 acpi_desc->idts.prev); acpi_nfit_init() 1655 list_cut_position(&prev.flushes, &acpi_desc->flushes, acpi_nfit_init() 1656 acpi_desc->flushes.prev); acpi_nfit_init() 1661 data = add_table(acpi_desc, &prev, data, end); acpi_nfit_init() 1670 rc = acpi_nfit_check_deletions(acpi_desc, &prev); acpi_nfit_init() 232 add_spa(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_system_address *spa) add_spa() argument 259 add_memdev(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_memory_map *memdev) add_memdev() argument 285 add_dcr(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_control_region *dcr) add_dcr() argument 310 add_bdw(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_data_region *bdw) add_bdw() argument 335 add_idt(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_interleave *idt) add_idt() argument 360 add_flush(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, struct acpi_nfit_flush_address *flush) add_flush() argument 385 add_table(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev, void *table, const void *end) add_table() argument 1611 acpi_nfit_check_deletions(struct acpi_nfit_desc *acpi_desc, struct nfit_table_prev *prev) acpi_nfit_check_deletions() argument
|
/linux-4.4.14/fs/nfs/ |
H A D | pagelist.c | 269 * @prev - the previous request in page group, or NULL if @req is the first 273 nfs_page_group_init(struct nfs_page *req, struct nfs_page *prev) nfs_page_group_init() argument 276 WARN_ON_ONCE(prev == req); nfs_page_group_init() 278 if (!prev) { nfs_page_group_init() 284 WARN_ON_ONCE(prev->wb_this_page != prev->wb_head); nfs_page_group_init() 285 WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &prev->wb_head->wb_flags)); nfs_page_group_init() 286 req->wb_head = prev->wb_head; nfs_page_group_init() 287 req->wb_this_page = prev->wb_this_page; nfs_page_group_init() 288 prev->wb_this_page = req; nfs_page_group_init() 297 if (test_bit(PG_INODE_REF, &prev->wb_head->wb_flags)) { nfs_page_group_init() 488 * @prev: previous request in desc, or NULL 495 struct nfs_page *prev, struct nfs_page *req) nfs_generic_pg_test() 921 * @prev: pointer to nfs_page 924 * The nfs_page structures 'prev' and 'req' are compared to ensure that the 930 static bool nfs_can_coalesce_requests(struct nfs_page *prev, nfs_can_coalesce_requests() argument 937 if (prev) { nfs_can_coalesce_requests() 938 if (!nfs_match_open_context(req->wb_context, prev->wb_context)) nfs_can_coalesce_requests() 945 prev->wb_lock_context)) nfs_can_coalesce_requests() 947 if (req_offset(req) != req_offset(prev) + prev->wb_bytes) nfs_can_coalesce_requests() 949 if (req->wb_page == prev->wb_page) { nfs_can_coalesce_requests() 950 if (req->wb_pgbase != prev->wb_pgbase + prev->wb_bytes) nfs_can_coalesce_requests() 954 prev->wb_pgbase + prev->wb_bytes != PAGE_CACHE_SIZE) nfs_can_coalesce_requests() 958 size = pgio->pg_ops->pg_test(pgio, prev, req); nfs_can_coalesce_requests() 978 struct nfs_page *prev = NULL; nfs_pageio_do_add_request() local 981 prev = nfs_list_entry(mirror->pg_list.prev); nfs_pageio_do_add_request() 987 if (!nfs_can_coalesce_requests(prev, req, desc)) nfs_pageio_do_add_request() 1271 struct nfs_page *prev; nfs_pageio_cond_complete() local 1277 prev = nfs_list_entry(mirror->pg_list.prev); nfs_pageio_cond_complete() 1278 if (index != prev->wb_index + 1) nfs_pageio_cond_complete() 494 nfs_generic_pg_test(struct nfs_pageio_descriptor *desc, struct nfs_page *prev, struct nfs_page *req) nfs_generic_pg_test() argument
|
/linux-4.4.14/fs/hfs/ |
H A D | bnode.c | 144 be32_to_cpu(desc.next), be32_to_cpu(desc.prev), hfs_bnode_dump() 179 if (node->prev) { hfs_bnode_unlink() 180 tmp = hfs_bnode_find(tree, node->prev); hfs_bnode_unlink() 194 tmp->prev = node->prev; hfs_bnode_unlink() 195 cnid = cpu_to_be32(tmp->prev); hfs_bnode_unlink() 196 hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, prev), 4); hfs_bnode_unlink() 199 tree->leaf_tail = node->prev; hfs_bnode_unlink() 202 if (!node->prev && !node->next) { hfs_bnode_unlink() 342 node->prev = be32_to_cpu(desc->prev); hfs_bnode_find()
|
/linux-4.4.14/sound/core/seq/ |
H A D | seq_prioq.c | 149 struct snd_seq_event_cell *cur, *prev; snd_seq_prioq_cell_in() local 179 prev = NULL; /* previous cell */ snd_seq_prioq_cell_in() 194 prev = cur; snd_seq_prioq_cell_in() 204 if (prev != NULL) snd_seq_prioq_cell_in() 205 prev->next = cell; snd_seq_prioq_cell_in() 294 struct snd_seq_event_cell *prev = NULL; snd_seq_prioq_leave() local 307 prev->next = cell->next; snd_seq_prioq_leave() 329 prev = cell; snd_seq_prioq_leave() 404 struct snd_seq_event_cell *prev = NULL; snd_seq_prioq_remove_events() local 420 prev->next = cell->next; snd_seq_prioq_remove_events() 437 prev = cell; snd_seq_prioq_remove_events()
|
/linux-4.4.14/fs/notify/dnotify/ |
H A D | dnotify.c | 93 struct dnotify_struct **prev; dnotify_handle_event() local 106 prev = &dn_mark->dn; dnotify_handle_event() 107 while ((dn = *prev) != NULL) { dnotify_handle_event() 109 prev = &dn->dn_next; dnotify_handle_event() 115 prev = &dn->dn_next; dnotify_handle_event() 117 *prev = dn->dn_next; dnotify_handle_event() 155 struct dnotify_struct **prev; dnotify_flush() local 171 prev = &dn_mark->dn; dnotify_flush() 172 while ((dn = *prev) != NULL) { dnotify_flush() 174 *prev = dn->dn_next; dnotify_flush() 179 prev = &dn->dn_next; dnotify_flush()
|
/linux-4.4.14/drivers/misc/mic/scif/ |
H A D | scif_rma_list.c | 31 struct scif_window *prev = list_entry(head, struct scif_window, list); scif_insert_tcw() local 37 curr = list_entry(head->prev, struct scif_window, list); scif_insert_tcw() 47 prev = curr; list_for_each() 49 list_add(&window->list, &prev->list); 60 struct scif_window *curr = NULL, *prev = NULL; scif_insert_window() local 68 prev = curr; list_for_each() 70 if (!prev) 73 list_add(&window->list, &prev->list);
|
/linux-4.4.14/drivers/mtd/ |
H A D | redboot.c | 180 struct fis_list *new_fl, **prev; parse_redboot_partitions() local 207 prev = &fl; parse_redboot_partitions() 208 while(*prev && (*prev)->img->flash_base < new_fl->img->flash_base) parse_redboot_partitions() 209 prev = &(*prev)->next; parse_redboot_partitions() 210 new_fl->next = *prev; parse_redboot_partitions() 211 *prev = new_fl; parse_redboot_partitions()
|
/linux-4.4.14/drivers/dma/ |
H A D | txx9dmac.c | 174 return list_entry(dc->active_list.prev, txx9dmac_last_active() 186 desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); txx9dmac_last_child() 435 struct txx9dmac_desc *prev = NULL; txx9dmac_dequeue() local 440 if (prev) { txx9dmac_dequeue() 441 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_dequeue() 443 prev->txd.phys, ddev->descsize, txx9dmac_dequeue() 446 prev = txx9dmac_last_child(desc); txx9dmac_dequeue() 723 struct txx9dmac_desc *prev; txx9dmac_prep_dma_memcpy() local 735 prev = first = NULL; txx9dmac_prep_dma_memcpy() 785 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_prep_dma_memcpy() 787 prev->txd.phys, ddev->descsize, txx9dmac_prep_dma_memcpy() 791 prev = desc; txx9dmac_prep_dma_memcpy() 796 txx9dmac_desc_set_INTENT(ddev, prev); txx9dmac_prep_dma_memcpy() 798 desc_write_CHAR(dc, prev, 0); txx9dmac_prep_dma_memcpy() 800 prev->txd.phys, ddev->descsize, txx9dmac_prep_dma_memcpy() 817 struct txx9dmac_desc *prev; txx9dmac_prep_slave_sg() local 832 prev = first = NULL; txx9dmac_prep_slave_sg() 879 desc_write_CHAR(dc, prev, desc->txd.phys); for_each_sg() 881 prev->txd.phys, for_each_sg() 886 prev = desc; for_each_sg() 891 txx9dmac_desc_set_INTENT(ddev, prev); 893 desc_write_CHAR(dc, prev, 0); 895 prev->txd.phys, ddev->descsize, 947 struct txx9dmac_desc *prev) txx9dmac_chain_dynamic() 953 prev = txx9dmac_last_child(prev); txx9dmac_chain_dynamic() 956 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_chain_dynamic() 958 prev->txd.phys, ddev->descsize, txx9dmac_chain_dynamic() 962 channel_read_CHAR(dc) == prev->txd.phys) txx9dmac_chain_dynamic() 981 struct txx9dmac_desc *prev = txx9dmac_last_active(dc); txx9dmac_issue_pending() local 983 if (!(prev->txd.flags & DMA_PREP_INTERRUPT) || txx9dmac_issue_pending() 985 txx9dmac_chain_dynamic(dc, prev); txx9dmac_issue_pending() 946 txx9dmac_chain_dynamic(struct txx9dmac_chan *dc, struct txx9dmac_desc *prev) txx9dmac_chain_dynamic() argument
|
H A D | img-mdc-dma.c | 294 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_dma_memcpy() local 313 if (prev) { mdc_prep_dma_memcpy() 314 prev->node_addr = curr_phys; mdc_prep_dma_memcpy() 315 prev->next_desc = curr; mdc_prep_dma_memcpy() 326 prev = curr; mdc_prep_dma_memcpy() 377 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_dma_cyclic() local 409 if (!prev) { mdc_prep_dma_cyclic() 413 prev->node_addr = curr_phys; mdc_prep_dma_cyclic() 414 prev->next_desc = curr; mdc_prep_dma_cyclic() 432 prev = curr; mdc_prep_dma_cyclic() 441 prev->node_addr = mdesc->list_phys; mdc_prep_dma_cyclic() 460 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_slave_sg() local 490 if (!prev) { for_each_sg() 494 prev->node_addr = curr_phys; for_each_sg() 495 prev->next_desc = curr; for_each_sg() 511 prev = curr; for_each_sg()
|
/linux-4.4.14/arch/arc/kernel/ |
H A D | fpu.c | 33 void fpu_save_restore(struct task_struct *prev, struct task_struct *next) fpu_save_restore() argument 35 unsigned int *saveto = &prev->thread.fpu.aux_dpfp[0].l; fpu_save_restore()
|
H A D | ctx_sw.c | 26 unsigned int prev = (unsigned int)prev_task; __switch_to() local 103 /* last (ret value) = prev : although for ARC it mov r0, r0 */ __switch_to() 109 : "n"(KSP_WORD_OFF), "r"(next), "r"(prev) __switch_to()
|
/linux-4.4.14/arch/c6x/kernel/ |
H A D | switch_to.S | 16 * void __switch_to(struct thread_info *prev, 22 || MV .L2X A4,B5 ; prev
|
/linux-4.4.14/kernel/trace/ |
H A D | trace_sched_switch.c | 20 struct task_struct *prev, struct task_struct *next) probe_sched_switch() 25 tracing_record_cmdline(prev); probe_sched_switch() 19 probe_sched_switch(void *ignore, bool preempt, struct task_struct *prev, struct task_struct *next) probe_sched_switch() argument
|
H A D | trace_stat.h | 16 void *(*stat_next)(void *prev, int idx);
|
/linux-4.4.14/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 38 struct sk_buff *prev, struct net_device *ldev); 132 struct sk_buff *prev, *next; lowpan_frag_queue() local 165 prev = fq->q.fragments_tail; lowpan_frag_queue() 166 if (!prev || lowpan_frag_queue() 167 lowpan_802154_cb(prev)->d_offset < lowpan_frag_queue() 172 prev = NULL; lowpan_frag_queue() 177 prev = next; lowpan_frag_queue() 185 if (prev) lowpan_frag_queue() 186 prev->next = skb; lowpan_frag_queue() 207 res = lowpan_frag_reasm(fq, prev, ldev); lowpan_frag_queue() 226 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *prev, lowpan_frag_reasm() argument 235 if (prev) { lowpan_frag_reasm() 236 head = prev->next; lowpan_frag_reasm() 245 prev->next = fp; lowpan_frag_reasm()
|
/linux-4.4.14/arch/sparc/kernel/ |
H A D | sun4m_smp.c | 123 int *prev; smp4m_smp_done() local 127 prev = &first; for_each_online_cpu() 129 *prev = i; for_each_online_cpu() 130 prev = &cpu_data(i).next; for_each_online_cpu() 132 *prev = first;
|
/linux-4.4.14/drivers/md/ |
H A D | dm-cache-policy-smq.c | 39 unsigned prev:28; member in struct:entry 137 return to_entry(es, e->prev); l_prev() 150 e->prev = INDEXER_NULL; l_add_head() 153 head->prev = l->head = to_index(es, e); l_add_head() 166 e->prev = l->tail; l_add_tail() 180 struct entry *prev = l_prev(es, old); l_add_before() local 182 if (!prev) l_add_before() 186 e->prev = old->prev; l_add_before() 188 prev->next = old->prev = to_index(es, e); l_add_before() 197 struct entry *prev = l_prev(es, e); l_del() local 200 if (prev) l_del() 201 prev->next = e->next; l_del() 206 next->prev = e->prev; l_del() 208 l->tail = e->prev; l_del() 609 struct entry **prev) __h_lookup() 613 *prev = NULL; __h_lookup() 618 *prev = e; __h_lookup() 625 struct entry *e, struct entry *prev) __h_unlink() 627 if (prev) __h_unlink() 628 prev->hash_next = e->hash_next; __h_unlink() 638 struct entry *e, *prev; h_lookup() local 641 e = __h_lookup(ht, h, oblock, &prev); h_lookup() 642 if (e && prev) { h_lookup() 647 __h_unlink(ht, h, e, prev); h_lookup() 657 struct entry *prev; h_remove() local 663 e = __h_lookup(ht, h, e->oblock, &prev); h_remove() 665 __h_unlink(ht, h, e, prev); h_remove() 700 e->prev = INDEXER_NULL; init_entry() 608 __h_lookup(struct hash_table *ht, unsigned h, dm_oblock_t oblock, struct entry **prev) __h_lookup() argument 624 __h_unlink(struct hash_table *ht, unsigned h, struct entry *e, struct entry *prev) __h_unlink() argument
|
/linux-4.4.14/fs/gfs2/ |
H A D | xattr.c | 77 struct gfs2_ea_header *prev, void *private); 82 struct gfs2_ea_header *ea, *prev = NULL; ea_foreach_i() local 88 for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { ea_foreach_i() 97 error = ea_call(ip, bh, ea, prev, data); ea_foreach_i() 167 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_find_i() 182 el->el_prev = prev; ea_find_i() 215 * @prev: 227 struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() 297 if (prev && !leave) { ea_dealloc_unstuffed() 300 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_dealloc_unstuffed() 301 prev->ea_rec_len = cpu_to_be32(len); ea_dealloc_unstuffed() 304 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_dealloc_unstuffed() 327 struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() 339 error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); ea_remove_unstuffed() 366 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_list_i() 838 struct gfs2_ea_header *prev = el->el_prev; ea_set_remove_stuffed() local 843 if (!prev || !GFS2_EA_IS_STUFFED(ea)) { ea_set_remove_stuffed() 846 } else if (GFS2_EA2NEXT(prev) != ea) { ea_set_remove_stuffed() 847 prev = GFS2_EA2NEXT(prev); ea_set_remove_stuffed() 848 gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea); ea_set_remove_stuffed() 851 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_set_remove_stuffed() 852 prev->ea_rec_len = cpu_to_be32(len); ea_set_remove_stuffed() 855 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_set_remove_stuffed() 924 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_set_simple() 939 error = ea_remove_unstuffed(ip, bh, ea, prev, 1); ea_set_simple() 1091 struct gfs2_ea_header *prev = el->el_prev; ea_remove_stuffed() local 1101 if (prev) { ea_remove_stuffed() 1104 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_remove_stuffed() 1105 prev->ea_rec_len = cpu_to_be32(len); ea_remove_stuffed() 1108 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_remove_stuffed() 166 ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_find_i() argument 225 ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() argument 325 ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() argument 365 ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_list_i() argument 923 ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_set_simple() argument
|
/linux-4.4.14/drivers/usb/host/whci/ |
H A D | asl.c | 29 struct whc_qset **next, struct whc_qset **prev) qset_get_next_prev() 38 p = qset->list_node.prev; qset_get_next_prev() 40 p = p->prev; qset_get_next_prev() 43 *prev = container_of(p, struct whc_qset, list_node); qset_get_next_prev() 55 struct whc_qset *next, *prev; asl_qset_insert() local 60 qset_get_next_prev(whc, qset, &next, &prev); asl_qset_insert() 62 whc_qset_set_link_ptr(&prev->qh.link, qset->qset_dma); asl_qset_insert() 68 struct whc_qset *prev, *next; asl_qset_remove() local 70 qset_get_next_prev(whc, qset, &next, &prev); asl_qset_remove() 83 whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma); asl_qset_remove() 28 qset_get_next_prev(struct whc *whc, struct whc_qset *qset, struct whc_qset **next, struct whc_qset **prev) qset_get_next_prev() argument
|
/linux-4.4.14/net/ipv4/ |
H A D | ip_fragment.c | 98 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, 336 struct sk_buff *prev, *next; ip_frag_queue() local 403 prev = qp->q.fragments_tail; ip_frag_queue() 404 if (!prev || FRAG_CB(prev)->offset < offset) { ip_frag_queue() 408 prev = NULL; ip_frag_queue() 412 prev = next; ip_frag_queue() 420 if (prev) { ip_frag_queue() 421 int i = (FRAG_CB(prev)->offset + prev->len) - offset; ip_frag_queue() 460 if (prev) ip_frag_queue() 461 prev->next = next; ip_frag_queue() 477 if (prev) ip_frag_queue() 478 prev->next = skb; ip_frag_queue() 508 err = ip_frag_reasm(qp, prev, dev); ip_frag_queue() 524 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, ip_frag_reasm() argument 543 if (prev) { ip_frag_reasm() 544 head = prev->next; ip_frag_reasm() 552 prev->next = fp; ip_frag_reasm()
|
/linux-4.4.14/drivers/staging/lustre/lustre/fld/ |
H A D | fld_cache.c | 226 curr = cache->fci_lru.prev; fld_cache_shrink() 232 curr = curr->prev; fld_cache_shrink() 308 * prev range only. fixup will handle overlapping with next range. */ fld_cache_overlap_handle() 343 fld_cache_entry_add(cache, f_new, f_curr->fce_list.prev); fld_cache_overlap_handle() 386 struct list_head *prev = NULL; fld_cache_insert_nolock() local 409 prev = &f_curr->fce_list; list_for_each_entry_safe() 418 if (prev == NULL) 419 prev = head; 423 fld_cache_entry_add(cache, f_new, prev); 520 struct fld_cache_entry *prev = NULL; fld_cache_lookup() local 529 if (prev != NULL) list_for_each_entry() 530 *range = prev->fce_range; list_for_each_entry() 534 prev = flde; list_for_each_entry()
|
/linux-4.4.14/drivers/iommu/ |
H A D | iova.c | 105 struct rb_node *prev, *curr = NULL; __alloc_and_insert_iova_range() local 114 prev = curr; __alloc_and_insert_iova_range() 131 prev = curr; __alloc_and_insert_iova_range() 153 /* If we have 'prev', it's a valid place to start the __alloc_and_insert_iova_range() 155 if (prev) __alloc_and_insert_iova_range() 156 entry = &prev; __alloc_and_insert_iova_range() 517 struct iova *prev = NULL, *next = NULL; split_and_remove_iova() local 521 prev = alloc_and_init_iova(iova->pfn_lo, pfn_lo - 1); split_and_remove_iova() 522 if (prev == NULL) split_and_remove_iova() 534 if (prev) { split_and_remove_iova() 535 iova_insert_rbtree(&iovad->rbroot, prev); split_and_remove_iova() 548 if (prev) split_and_remove_iova() 549 free_iova_mem(prev); split_and_remove_iova()
|
/linux-4.4.14/arch/cris/arch-v10/kernel/ |
H A D | fasttimer.c | 260 t->prev = NULL; start_one_shot_timer() 264 fast_timer_list->prev = t; start_one_shot_timer() 279 t->prev = tmp; start_one_shot_timer() 283 tmp->next->prev = t; start_one_shot_timer() 296 return (t->next != NULL) || (t->prev != NULL) || (t == fast_timer_list); fast_timer_pending() 301 struct fast_timer *next, *prev; detach_fast_timer() local 305 prev = t->prev; detach_fast_timer() 307 next->prev = prev; detach_fast_timer() 308 if (prev) detach_fast_timer() 309 prev->next = next; detach_fast_timer() 323 t->next = t->prev = NULL; del_fast_timer() 384 if (t->prev) timer1_handler() 386 t->prev->next = t->next; timer1_handler() 394 t->next->prev = t->prev; timer1_handler() 396 t->prev = NULL; timer1_handler()
|
/linux-4.4.14/arch/cris/arch-v32/kernel/ |
H A D | fasttimer.c | 238 t->prev = NULL; start_one_shot_timer() 241 fast_timer_list->prev = t; start_one_shot_timer() 253 t->prev = tmp; start_one_shot_timer() 257 tmp->next->prev = t; start_one_shot_timer() 270 return (t->next != NULL) || (t->prev != NULL) || (t == fast_timer_list); fast_timer_pending() 275 struct fast_timer *next, *prev; detach_fast_timer() local 279 prev = t->prev; detach_fast_timer() 281 next->prev = prev; detach_fast_timer() 282 if (prev) detach_fast_timer() 283 prev->next = next; detach_fast_timer() 297 t->next = t->prev = NULL; del_fast_timer() 373 if (t->prev) timer_trig_handler() 374 t->prev->next = t->next; timer_trig_handler() 378 t->next->prev = t->prev; timer_trig_handler() 379 t->prev = NULL; timer_trig_handler()
|
/linux-4.4.14/drivers/mmc/core/ |
H A D | sdio_cis.c | 231 struct sdio_func_tuple *this, **prev; sdio_read_cis() local 255 prev = &func->tuples; sdio_read_cis() 257 prev = &card->tuples; sdio_read_cis() 259 BUG_ON(*prev); sdio_read_cis() 311 *prev = this; sdio_read_cis() 312 prev = &this->next; sdio_read_cis() 341 *prev = card->tuples; sdio_read_cis()
|
/linux-4.4.14/arch/sh/kernel/cpu/sh5/ |
H A D | switchto.S | 24 r2 - prev 25 r3 - &prev->thread 30 r2 - last (=prev) : this just stays in r2 throughout 120 st.l r3, 0, r15 ! prev->thread.sp 121 st.l r3, 8, r1 ! prev->thread.kregs 122 st.l r3, 4, r9 ! prev->thread.pc 123 st.q r1, 0, r9 ! save prev->thread.pc into pt_regs->pc
|
/linux-4.4.14/arch/m68k/sun3/ |
H A D | sun3dvma.c | 110 struct hole *prev = NULL; refill() local 117 if(!prev) { refill() 118 prev = hole; refill() 122 if(hole->end == prev->start) { refill() 123 hole->size += prev->size; refill() 124 hole->end = prev->end; refill() 125 list_move(&(prev->list), &hole_cache); refill()
|
/linux-4.4.14/net/atm/ |
H A D | mpoa_caches.c | 112 entry->prev = NULL; in_cache_add_entry() 114 client->in_cache->prev = entry; in_cache_add_entry() 200 if (entry->prev != NULL) in_cache_remove_entry() 201 entry->prev->next = entry->next; in_cache_remove_entry() 205 entry->next->prev = entry->prev; in_cache_remove_entry() 435 if (entry->prev != NULL) eg_cache_remove_entry() 436 entry->prev->next = entry->next; eg_cache_remove_entry() 440 entry->next->prev = entry->prev; eg_cache_remove_entry() 475 entry->prev = NULL; eg_cache_add_entry() 477 client->eg_cache->prev = entry; eg_cache_add_entry()
|
H A D | mpoa_caches.h | 16 struct in_cache_entry *prev; member in struct:in_cache_entry 53 struct eg_cache_entry *prev; member in struct:eg_cache_entry
|
/linux-4.4.14/fs/autofs4/ |
H A D | expire.c | 76 static struct dentry *get_next_positive_subdir(struct dentry *prev, get_next_positive_subdir() argument 86 if (prev) get_next_positive_subdir() 87 next = prev->d_child.next; get_next_positive_subdir() 89 prev = dget_dlock(root); get_next_positive_subdir() 90 next = prev->d_subdirs.next; get_next_positive_subdir() 97 dput(prev); get_next_positive_subdir() 115 dput(prev); get_next_positive_subdir() 123 static struct dentry *get_next_positive_dentry(struct dentry *prev, get_next_positive_dentry() argument 130 if (prev == NULL) get_next_positive_dentry() 135 p = prev; get_next_positive_dentry() 146 dput(prev); get_next_positive_dentry() 178 dput(prev); get_next_positive_dentry()
|
/linux-4.4.14/kernel/gcov/ |
H A D | base.c | 137 struct gcov_info *prev = NULL; gcov_module_notifier() local 146 gcov_info_unlink(prev, info); gcov_module_notifier() 150 prev = info; gcov_module_notifier()
|
/linux-4.4.14/net/ipv6/ |
H A D | reassembly.c | 79 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, 212 struct sk_buff *prev, *next; ip6_frag_queue() local 289 prev = fq->q.fragments_tail; ip6_frag_queue() 290 if (!prev || FRAG6_CB(prev)->offset < offset) { ip6_frag_queue() 294 prev = NULL; ip6_frag_queue() 298 prev = next; ip6_frag_queue() 310 if (prev && ip6_frag_queue() 311 (FRAG6_CB(prev)->offset + prev->len) > offset) ip6_frag_queue() 324 if (prev) ip6_frag_queue() 325 prev->next = skb; ip6_frag_queue() 353 res = ip6_frag_reasm(fq, prev, dev); ip6_frag_queue() 379 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, ip6_frag_reasm() argument 396 if (prev) { ip6_frag_reasm() 397 head = prev->next; ip6_frag_reasm() 406 prev->next = fp; ip6_frag_reasm()
|
/linux-4.4.14/drivers/md/bcache/ |
H A D | bset.c | 280 /* Space required for the prev pointers */ bset_prev_bytes() 293 kfree(t->prev); bch_btree_keys_free() 295 free_pages((unsigned long) t->prev, bch_btree_keys_free() 306 t->prev = NULL; bch_btree_keys_free() 330 t->prev = bset_prev_bytes(b) < PAGE_SIZE bch_btree_keys_alloc() 333 if (!t->prev) bch_btree_keys_alloc() 499 * bits we're going to store in bkey_float->mantissa. t->prev[j] stores the size 528 return (void *) (((uint64_t *) tree_to_bkey(t, j)) - t->prev[j]); tree_to_prev_bkey() 533 * maintain a full search tree, we just keep a simple lookup table in t->prev. 537 return cacheline_to_bkey(t, cacheline, t->prev[cacheline]); table_to_bkey() 596 t->prev = t[-1].prev + j; bset_alloc_tree() 613 t->prev[0] = bkey_to_cacheline_offset(t, 0, t->data->start); bch_bset_build_unwritten_tree() 637 struct bkey *prev = NULL, *k = t->data->start; bch_bset_build_written_tree() local 660 prev = k, k = bkey_next(k); bch_bset_build_written_tree() 662 t->prev[j] = bkey_u64s(prev); bch_bset_build_written_tree() 750 t->prev[j] += shift; bch_bset_fix_lookup_table() 752 if (t->prev[j] > 7) { bch_bset_fix_lookup_table() 758 t->prev[j] = bkey_to_cacheline_offset(t, j, k); bch_bset_fix_lookup_table() 771 t->prev[t->size] = bkey_to_cacheline_offset(t, t->size, k); bch_bset_fix_lookup_table() 824 struct bkey *m, *prev = NULL; bch_btree_insert_key() local 840 prev = m, m = bkey_next(m); bch_btree_insert_key() 842 /* prev is in the tree, if we merge we're done */ bch_btree_insert_key() 844 if (prev && bch_btree_insert_key() 845 bch_bkey_try_merge(b, prev, k)) bch_btree_insert_key()
|
/linux-4.4.14/drivers/scsi/ |
H A D | scsi_module.c | 47 while ((l = l->prev) != &sht->legacy_hosts) init_this_scsi_driver()
|
/linux-4.4.14/drivers/media/pci/cx88/ |
H A D | cx88-vbi.c | 168 struct cx88_buffer *prev; buffer_queue() local 184 prev = list_entry(q->active.prev, struct cx88_buffer, list); buffer_queue() 186 prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); buffer_queue()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/bios/ |
H A D | base.c | 73 u8 *prev = bios->data; nvbios_extend() local 75 bios->data = prev; nvbios_extend() 78 memcpy(bios->data, prev, bios->size); nvbios_extend() 80 kfree(prev); nvbios_extend()
|
/linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/volt/ |
H A D | base.c | 100 int prev = nvkm_volt_get(volt); nvkm_volt_set_id() local 101 if (!condition || prev < 0 || nvkm_volt_set_id() 102 (condition < 0 && ret < prev) || nvkm_volt_set_id() 103 (condition > 0 && ret > prev)) { nvkm_volt_set_id()
|
/linux-4.4.14/fs/ |
H A D | userfaultfd.c | 396 * list_del_init() to refile across the two lists, the prev handle_userfault() 430 struct vm_area_struct *vma, *prev; userfaultfd_release() local 446 prev = NULL; userfaultfd_release() 452 prev = vma; userfaultfd_release() 456 prev = vma_merge(mm, prev, vma->vm_start, vma->vm_end, userfaultfd_release() 461 if (prev) userfaultfd_release() 462 vma = prev; userfaultfd_release() 464 prev = vma; userfaultfd_release() 721 struct vm_area_struct *vma, *prev, *cur; userfaultfd_register() local 764 vma = find_vma_prev(mm, start, &prev); userfaultfd_register() 810 prev = vma; userfaultfd_register() 833 prev = vma_merge(mm, prev, start, vma_end, new_flags, userfaultfd_register() 837 if (prev) { userfaultfd_register() 838 vma = prev; userfaultfd_register() 861 prev = vma; userfaultfd_register() 885 struct vm_area_struct *vma, *prev, *cur; userfaultfd_unregister() local 906 vma = find_vma_prev(mm, start, &prev); userfaultfd_unregister() 947 prev = vma; userfaultfd_unregister() 967 prev = vma_merge(mm, prev, start, vma_end, new_flags, userfaultfd_unregister() 971 if (prev) { userfaultfd_unregister() 972 vma = prev; userfaultfd_unregister() 995 prev = vma; userfaultfd_unregister()
|
/linux-4.4.14/arch/sh/kernel/ |
H A D | process_32.c | 181 __switch_to(struct task_struct *prev, struct task_struct *next) __switch_to() argument 189 unlazy_fpu(prev, task_pt_regs(prev)); __switch_to() 213 return prev; __switch_to()
|
/linux-4.4.14/arch/arm64/include/asm/ |
H A D | mmu_context.h | 127 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 132 if (prev == next) switch_mm() 148 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
|
/linux-4.4.14/fs/hfsplus/ |
H A D | bnode.c | 308 be32_to_cpu(desc.next), be32_to_cpu(desc.prev), hfs_bnode_dump() 343 if (node->prev) { hfs_bnode_unlink() 344 tmp = hfs_bnode_find(tree, node->prev); hfs_bnode_unlink() 359 tmp->prev = node->prev; hfs_bnode_unlink() 360 cnid = cpu_to_be32(tmp->prev); hfs_bnode_unlink() 362 offsetof(struct hfs_bnode_desc, prev), 4); hfs_bnode_unlink() 365 tree->leaf_tail = node->prev; hfs_bnode_unlink() 368 if (!node->prev && !node->next) hfs_bnode_unlink() 510 node->prev = be32_to_cpu(desc->prev); hfs_bnode_find()
|
/linux-4.4.14/net/core/ |
H A D | dst.c | 322 unsigned long prev, new; dst_cow_metrics_generic() local 327 prev = cmpxchg(&dst->_metrics, old, new); dst_cow_metrics_generic() 329 if (prev != old) { dst_cow_metrics_generic() 331 p = __DST_METRICS_PTR(prev); dst_cow_metrics_generic() 332 if (prev & DST_METRICS_READ_ONLY) dst_cow_metrics_generic() 343 unsigned long prev, new; __dst_destroy_metrics_generic() local 346 prev = cmpxchg(&dst->_metrics, old, new); __dst_destroy_metrics_generic() 347 if (prev == old) __dst_destroy_metrics_generic()
|
/linux-4.4.14/tools/perf/util/ |
H A D | cpumap.c | 52 int n, cpu, prev; cpu_map__read() local 56 prev = -1; cpu_map__read() 61 if (prev >= 0) { cpu_map__read() 62 int new_max = nr_cpus + cpu - prev - 1; cpu_map__read() 72 while (++prev < cpu) cpu_map__read() 73 tmp_cpus[nr_cpus++] = prev; cpu_map__read() 85 prev = cpu; cpu_map__read() 87 prev = -1; cpu_map__read()
|
/linux-4.4.14/fs/nilfs2/ |
H A D | alloc.c | 212 struct nilfs_bh_assoc *prev, nilfs_palloc_get_block() 218 if (prev->bh && blkoff == prev->blkoff) { nilfs_palloc_get_block() 219 get_bh(prev->bh); nilfs_palloc_get_block() 220 *bhp = prev->bh; nilfs_palloc_get_block() 233 brelse(prev->bh); nilfs_palloc_get_block() 235 prev->bh = *bhp; nilfs_palloc_get_block() 236 prev->blkoff = blkoff; nilfs_palloc_get_block() 246 * @prev: nilfs_bh_assoc struct of the last used buffer 247 * @lock: spin lock protecting @prev 250 struct nilfs_bh_assoc *prev, nilfs_palloc_delete_block() 254 if (prev->bh && blkoff == prev->blkoff) { nilfs_palloc_delete_block() 255 brelse(prev->bh); nilfs_palloc_delete_block() 256 prev->bh = NULL; nilfs_palloc_delete_block() 206 nilfs_palloc_get_block(struct inode *inode, unsigned long blkoff, int create, void (*init_block)(struct inode *, struct buffer_head *, void *), struct buffer_head **bhp, struct nilfs_bh_assoc *prev, spinlock_t *lock) nilfs_palloc_get_block() argument 249 nilfs_palloc_delete_block(struct inode *inode, unsigned long blkoff, struct nilfs_bh_assoc *prev, spinlock_t *lock) nilfs_palloc_delete_block() argument
|
/linux-4.4.14/drivers/acpi/acpica/ |
H A D | psloop.c | 319 union acpi_operand_object *prev; acpi_ps_link_module_code() local 328 prev = next = acpi_gbl_module_code_list; acpi_ps_link_module_code() 330 prev = next; acpi_ps_link_module_code() 338 if (!prev || acpi_ps_link_module_code() 339 ((prev->method.aml_start + prev->method.aml_length) != aml_start)) { acpi_ps_link_module_code() 370 if (!prev) { acpi_ps_link_module_code() 373 prev->method.mutex = method_obj; acpi_ps_link_module_code() 378 prev)); acpi_ps_link_module_code() 380 prev->method.aml_length += aml_length; acpi_ps_link_module_code()
|
/linux-4.4.14/arch/x86/kernel/ |
H A D | process_64.c | 277 struct thread_struct *prev = &prev_p->thread; __switch_to() local 279 struct fpu *prev_fpu = &prev->fpu; __switch_to() 325 savesegment(es, prev->es); __switch_to() 326 if (unlikely(next->es | prev->es)) __switch_to() 329 savesegment(ds, prev->ds); __switch_to() 330 if (unlikely(next->ds | prev->ds)) __switch_to() 368 if (unlikely(fsindex | next->fsindex | prev->fs)) { __switch_to() 381 prev->fs = 0; __switch_to() 385 prev->fsindex = fsindex; __switch_to() 387 if (unlikely(gsindex | next->gsindex | prev->gs)) { __switch_to() 392 prev->gs = 0; __switch_to() 396 prev->gsindex = gsindex; __switch_to() 422 prev->iopl != next->iopl)) __switch_to()
|