/linux-4.1.27/lib/ |
H A D | list_debug.c | 19 * the prev/next entries already! 23 struct list_head *prev, __list_add() 26 WARN(next->prev != prev, __list_add() 27 "list_add corruption. next->prev should be " __list_add() 28 "prev (%p), but was %p. (next=%p).\n", __list_add() 29 prev, next->prev, next); __list_add() 30 WARN(prev->next != next, __list_add() 31 "list_add corruption. prev->next should be " __list_add() 32 "next (%p), but was %p. (prev=%p).\n", __list_add() 33 next, prev->next, prev); __list_add() 34 WARN(new == prev || new == next, __list_add() 35 "list_add double add: new=%p, prev=%p, next=%p.\n", __list_add() 36 new, prev, next); __list_add() 37 next->prev = new; __list_add() 39 new->prev = prev; __list_add() 40 prev->next = new; __list_add() 46 struct list_head *prev, *next; __list_del_entry() local 48 prev = entry->prev; __list_del_entry() 54 WARN(prev == LIST_POISON2, __list_del_entry() 55 "list_del corruption, %p->prev is LIST_POISON2 (%p)\n", __list_del_entry() 57 WARN(prev->next != entry, __list_del_entry() 58 "list_del corruption. prev->next should be %p, " __list_del_entry() 59 "but was %p\n", entry, prev->next) || __list_del_entry() 60 WARN(next->prev != entry, __list_del_entry() 61 "list_del corruption. next->prev should be %p, " __list_del_entry() 62 "but was %p\n", entry, next->prev)) __list_del_entry() 65 __list_del(prev, next); __list_del_entry() 79 entry->prev = LIST_POISON2; list_del() 87 struct list_head *prev, struct list_head *next) __list_add_rcu() 89 WARN(next->prev != prev, __list_add_rcu() 90 "list_add_rcu corruption. next->prev should be prev (%p), but was %p. (next=%p).\n", __list_add_rcu() 91 prev, next->prev, next); __list_add_rcu() 92 WARN(prev->next != next, __list_add_rcu() 93 "list_add_rcu corruption. prev->next should be next (%p), but was %p. (prev=%p).\n", __list_add_rcu() 94 next, prev->next, prev); __list_add_rcu() 96 new->prev = prev; __list_add_rcu() 97 rcu_assign_pointer(list_next_rcu(prev), new); __list_add_rcu() 98 next->prev = new; __list_add_rcu() 22 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 86 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument
|
H A D | plist.c | 36 WARN(n->prev != p || p->next != n, plist_check_prev_next() 38 "prev: %p, n: %p, p: %p\n" plist_check_prev_next() 40 t, t->next, t->prev, plist_check_prev_next() 41 p, p->next, p->prev, plist_check_prev_next() 42 n, n->next, n->prev); plist_check_prev_next() 47 struct list_head *prev = top, *next = top->next; plist_check_list() local 49 plist_check_prev_next(top, prev, next); plist_check_list() 51 prev = next; plist_check_list() 52 next = prev->next; plist_check_list() 53 plist_check_prev_next(top, prev, next); plist_check_list() 76 struct plist_node *first, *iter, *prev = NULL; plist_add() local 94 prev = iter; plist_add() 99 if (!prev || prev->prio != node->prio) plist_add()
|
H A D | list_sort.c | 17 * sentinel head node, "prev" links not maintained. 45 * prev-link restoration pass, or maintaining the prev links 61 a->prev = tail; merge_and_restore_back_links() 65 b->prev = tail; merge_and_restore_back_links() 82 tail->next->prev = tail; merge_and_restore_back_links() 87 head->prev = tail; merge_and_restore_back_links() 119 head->prev->next = NULL; list_sort() 249 if (cur->next->prev != cur) { list_sort_test() 274 if (head.prev != cur) { list_sort_test()
|
/linux-4.1.27/arch/score/include/asm/ |
H A D | switch_to.h | 6 #define switch_to(prev, next, last) \ 8 (last) = resume(prev, next, task_thread_info(next)); \ 11 #define finish_arch_switch(prev) do {} while (0)
|
H A D | mmu_context.h | 72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 102 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument
|
/linux-4.1.27/arch/parisc/include/asm/ |
H A D | switch_to.h | 8 #define switch_to(prev, next, last) do { \ 9 (last) = _switch_to(prev, next); \
|
H A D | mmu_context.h | 52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) switch_mm() argument 55 if (prev != next) { switch_mm() 63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 78 switch_mm(prev,next,current); activate_mm()
|
/linux-4.1.27/arch/arm/include/asm/ |
H A D | switch_to.h | 13 #define finish_arch_switch(prev) dsb(ish) 17 * switch_to(prev, next) should switch from task `prev' to `next' 18 * `prev' will never be the same as `next'. schedule() itself 23 #define switch_to(prev,next,last) \ 25 last = __switch_to(prev,task_thread_info(prev), task_thread_info(next)); \
|
H A D | mmu_context.h | 93 #define activate_mm(prev,next) switch_mm(prev, next, NULL) 116 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 132 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) { switch_mm() 135 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
|
/linux-4.1.27/tools/firewire/ |
H A D | list.h | 2 struct list *next, *prev; member in struct:list 9 list->prev = list; list_init() 21 new_link->prev = link->prev; list_insert() 23 new_link->prev->next = new_link; list_insert() 24 new_link->next->prev = new_link; list_insert() 42 link->prev->next = link->next; list_remove() 43 link->next->prev = link->prev; list_remove() 53 list_entry((list)->prev, type, member)
|
H A D | nosy-dump.c | 498 struct subaction *sa, *prev; handle_request_packet() local 507 prev = list_tail(&t->request_list, handle_request_packet() 510 if (!ACK_BUSY(prev->ack)) { handle_request_packet() 519 if (prev->packet.common.tcode != sa->packet.common.tcode || handle_request_packet() 520 prev->packet.common.tlabel != sa->packet.common.tlabel) { handle_request_packet() 563 struct subaction *sa, *prev; handle_response_packet() local 576 prev = list_tail(&t->response_list, struct subaction, link); handle_response_packet() 578 if (!ACK_BUSY(prev->ack)) { handle_response_packet() 585 if (prev->packet.common.tcode != sa->packet.common.tcode || handle_response_packet() 586 prev->packet.common.tlabel != sa->packet.common.tlabel) { handle_response_packet() 591 prev = list_tail(&t->request_list, struct subaction, link); handle_response_packet() 592 if (prev->ack != ACK_PENDING) { handle_response_packet() 599 if (packet_info[prev->packet.common.tcode].response_tcode != handle_response_packet()
|
/linux-4.1.27/include/asm-generic/ |
H A D | cmpxchg-local.h | 17 unsigned long flags, prev; __cmpxchg_local_generic() local 27 case 1: prev = *(u8 *)ptr; __cmpxchg_local_generic() 28 if (prev == old) __cmpxchg_local_generic() 31 case 2: prev = *(u16 *)ptr; __cmpxchg_local_generic() 32 if (prev == old) __cmpxchg_local_generic() 35 case 4: prev = *(u32 *)ptr; __cmpxchg_local_generic() 36 if (prev == old) __cmpxchg_local_generic() 39 case 8: prev = *(u64 *)ptr; __cmpxchg_local_generic() 40 if (prev == old) __cmpxchg_local_generic() 47 return prev; __cmpxchg_local_generic() 56 u64 prev; __cmpxchg64_local_generic() local 60 prev = *(u64 *)ptr; __cmpxchg64_local_generic() 61 if (prev == old) __cmpxchg64_local_generic() 64 return prev; __cmpxchg64_local_generic()
|
H A D | mutex-xchg.h | 94 int prev = atomic_xchg(count, 0); __mutex_fastpath_trylock() local 96 if (unlikely(prev < 0)) { __mutex_fastpath_trylock() 99 * state. If while doing so we get back a prev value of 1 __mutex_fastpath_trylock() 108 prev = atomic_xchg(count, prev); __mutex_fastpath_trylock() 109 if (prev < 0) __mutex_fastpath_trylock() 110 prev = 0; __mutex_fastpath_trylock() 113 return prev; __mutex_fastpath_trylock()
|
H A D | switch_to.h | 25 #define switch_to(prev, next, last) \ 27 ((last) = __switch_to((prev), (next))); \
|
H A D | mmu_context.h | 34 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument
|
/linux-4.1.27/arch/unicore32/include/asm/ |
H A D | switch_to.h | 17 * switch_to(prev, next) should switch from task `prev' to `next' 18 * `prev' will never be the same as `next'. schedule() itself 24 #define switch_to(prev, next, last) \ 26 last = __switch_to(prev, task_thread_info(prev), \
|
H A D | mmu_context.h | 49 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 54 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) switch_mm() 59 #define activate_mm(prev, next) switch_mm(prev, next, NULL)
|
/linux-4.1.27/arch/blackfin/include/asm/ |
H A D | switch_to.h | 21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next); 24 #define switch_to(prev,next,last) \ 26 memcpy (&task_thread_info(prev)->l1_task_info, L1_SCRATCH_TASK_INFO, \ 30 (last) = resume (prev, next); \ 33 #define switch_to(prev, next, last) \ 35 (last) = resume(prev, next); \
|
H A D | mmu_context.h | 71 #define activate_mm(prev, next) switch_mm(prev, next, NULL) 111 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 116 __switch_mm(prev, next, tsk); switch_mm() 160 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 163 __switch_mm(prev, next, tsk); switch_mm()
|
/linux-4.1.27/drivers/media/platform/omap3isp/ |
H A D | isppreview.c | 152 preview_config_luma_enhancement(struct isp_prev_device *prev, preview_config_luma_enhancement() argument 155 struct isp_device *isp = to_isp_device(prev); preview_config_luma_enhancement() 171 preview_enable_luma_enhancement(struct isp_prev_device *prev, bool enable) preview_enable_luma_enhancement() argument 173 struct isp_device *isp = to_isp_device(prev); preview_enable_luma_enhancement() 186 static void preview_enable_invalaw(struct isp_prev_device *prev, bool enable) preview_enable_invalaw() argument 188 struct isp_device *isp = to_isp_device(prev); preview_enable_invalaw() 201 static void preview_config_hmed(struct isp_prev_device *prev, preview_config_hmed() argument 204 struct isp_device *isp = to_isp_device(prev); preview_config_hmed() 216 static void preview_enable_hmed(struct isp_prev_device *prev, bool enable) preview_enable_hmed() argument 218 struct isp_device *isp = to_isp_device(prev); preview_enable_hmed() 236 static void preview_config_cfa(struct isp_prev_device *prev, preview_config_cfa() argument 245 const unsigned int *order = cfa_coef_order[prev->params.cfa_order]; preview_config_cfa() 247 struct isp_device *isp = to_isp_device(prev); preview_config_cfa() 272 preview_config_chroma_suppression(struct isp_prev_device *prev, preview_config_chroma_suppression() argument 275 struct isp_device *isp = to_isp_device(prev); preview_config_chroma_suppression() 288 preview_enable_chroma_suppression(struct isp_prev_device *prev, bool enable) preview_enable_chroma_suppression() argument 290 struct isp_device *isp = to_isp_device(prev); preview_enable_chroma_suppression() 306 preview_config_whitebalance(struct isp_prev_device *prev, preview_config_whitebalance() argument 309 struct isp_device *isp = to_isp_device(prev); preview_config_whitebalance() 345 preview_config_blkadj(struct isp_prev_device *prev, preview_config_blkadj() argument 348 struct isp_device *isp = to_isp_device(prev); preview_config_blkadj() 361 preview_config_rgb_blending(struct isp_prev_device *prev, preview_config_rgb_blending() argument 364 struct isp_device *isp = to_isp_device(prev); preview_config_rgb_blending() 399 preview_config_csc(struct isp_prev_device *prev, preview_config_csc() argument 402 struct isp_device *isp = to_isp_device(prev); preview_config_csc() 431 preview_config_yc_range(struct isp_prev_device *prev, preview_config_yc_range() argument 434 struct isp_device *isp = to_isp_device(prev); preview_config_yc_range() 449 preview_config_dcor(struct isp_prev_device *prev, preview_config_dcor() argument 452 struct isp_device *isp = to_isp_device(prev); preview_config_dcor() 471 static void preview_enable_dcor(struct isp_prev_device *prev, bool enable) preview_enable_dcor() argument 473 struct isp_device *isp = to_isp_device(prev); preview_enable_dcor() 487 preview_enable_drkframe_capture(struct isp_prev_device *prev, bool enable) preview_enable_drkframe_capture() argument 489 struct isp_device *isp = to_isp_device(prev); preview_enable_drkframe_capture() 502 static void preview_enable_drkframe(struct isp_prev_device *prev, bool enable) preview_enable_drkframe() argument 504 struct isp_device *isp = to_isp_device(prev); preview_enable_drkframe() 518 preview_config_noisefilter(struct isp_prev_device *prev, preview_config_noisefilter() argument 521 struct isp_device *isp = to_isp_device(prev); preview_config_noisefilter() 538 preview_enable_noisefilter(struct isp_prev_device *prev, bool enable) preview_enable_noisefilter() argument 540 struct isp_device *isp = to_isp_device(prev); preview_enable_noisefilter() 554 preview_config_gammacorrn(struct isp_prev_device *prev, preview_config_gammacorrn() argument 557 struct isp_device *isp = to_isp_device(prev); preview_config_gammacorrn() 587 preview_enable_gammacorrn(struct isp_prev_device *prev, bool enable) preview_enable_gammacorrn() argument 589 struct isp_device *isp = to_isp_device(prev); preview_enable_gammacorrn() 605 preview_config_contrast(struct isp_prev_device *prev, preview_config_contrast() argument 608 struct isp_device *isp = to_isp_device(prev); preview_config_contrast() 619 preview_config_brightness(struct isp_prev_device *prev, preview_config_brightness() argument 622 struct isp_device *isp = to_isp_device(prev); preview_config_brightness() 636 preview_update_contrast(struct isp_prev_device *prev, u8 contrast) preview_update_contrast() argument 641 spin_lock_irqsave(&prev->params.lock, flags); preview_update_contrast() 642 params = (prev->params.active & OMAP3ISP_PREV_CONTRAST) preview_update_contrast() 643 ? &prev->params.params[0] : &prev->params.params[1]; preview_update_contrast() 649 spin_unlock_irqrestore(&prev->params.lock, flags); preview_update_contrast() 658 preview_update_brightness(struct isp_prev_device *prev, u8 brightness) preview_update_brightness() argument 663 spin_lock_irqsave(&prev->params.lock, flags); preview_update_brightness() 664 params = (prev->params.active & OMAP3ISP_PREV_BRIGHTNESS) preview_update_brightness() 665 ? &prev->params.params[0] : &prev->params.params[1]; preview_update_brightness() 671 spin_unlock_irqrestore(&prev->params.lock, flags); preview_update_brightness() 675 preview_params_lock(struct isp_prev_device *prev, u32 update, bool shadow) preview_params_lock() argument 677 u32 active = prev->params.active; preview_params_lock() 681 prev->params.params[0].busy |= ~active & update; preview_params_lock() 682 prev->params.params[1].busy |= active & update; preview_params_lock() 685 update = (prev->params.params[0].update & active) preview_params_lock() 686 | (prev->params.params[1].update & ~active); preview_params_lock() 688 prev->params.params[0].busy |= active & update; preview_params_lock() 689 prev->params.params[1].busy |= ~active & update; preview_params_lock() 696 preview_params_unlock(struct isp_prev_device *prev, u32 update, bool shadow) preview_params_unlock() argument 698 u32 active = prev->params.active; preview_params_unlock() 704 prev->params.params[0].update |= (~active & update); preview_params_unlock() 705 prev->params.params[1].update |= (active & update); preview_params_unlock() 706 prev->params.params[0].busy &= active; preview_params_unlock() 707 prev->params.params[1].busy &= ~active; preview_params_unlock() 712 prev->params.params[0].update &= ~(active & update); preview_params_unlock() 713 prev->params.params[1].update &= ~(~active & update); preview_params_unlock() 714 prev->params.params[0].busy &= ~active; preview_params_unlock() 715 prev->params.params[1].busy &= active; preview_params_unlock() 719 static void preview_params_switch(struct isp_prev_device *prev) preview_params_switch() argument 727 to_switch = (prev->params.params[0].update & ~prev->params.active) preview_params_switch() 728 | (prev->params.params[1].update & prev->params.active); preview_params_switch() 729 to_switch &= ~(prev->params.params[0].busy | preview_params_switch() 730 prev->params.params[1].busy); preview_params_switch() 734 prev->params.active ^= to_switch; preview_params_switch() 739 prev->params.params[0].update &= ~(~prev->params.active & to_switch); preview_params_switch() 740 prev->params.params[1].update &= ~(prev->params.active & to_switch); preview_params_switch() 856 * @prev: ISP preview engine 862 static int preview_config(struct isp_prev_device *prev, preview_config() argument 875 spin_lock_irqsave(&prev->params.lock, flags); preview_config() 876 preview_params_lock(prev, cfg->update, true); preview_config() 877 active = prev->params.active; preview_config() 878 spin_unlock_irqrestore(&prev->params.lock, flags); preview_config() 890 params = &prev->params.params[!!(active & bit)]; preview_config() 912 spin_lock_irqsave(&prev->params.lock, flags); preview_config() 913 preview_params_unlock(prev, update, true); preview_config() 914 preview_params_switch(prev); preview_config() 915 spin_unlock_irqrestore(&prev->params.lock, flags); preview_config() 922 * @prev: pointer to preview private structure 928 static void preview_setup_hw(struct isp_prev_device *prev, u32 update, preview_setup_hw() argument 937 features = (prev->params.params[0].features & active) preview_setup_hw() 938 | (prev->params.params[1].features & ~active); preview_setup_hw() 948 params = &prev->params.params[!(active & bit)]; preview_setup_hw() 952 attr->config(prev, params); preview_setup_hw() 954 attr->enable(prev, true); preview_setup_hw() 957 attr->enable(prev, false); preview_setup_hw() 964 * @prev: pointer to previewer private structure 967 static void preview_config_ycpos(struct isp_prev_device *prev, u32 pixelcode) preview_config_ycpos() argument 969 struct isp_device *isp = to_isp_device(prev); preview_config_ycpos() 992 static void preview_config_averager(struct isp_prev_device *prev, u8 average) preview_config_averager() argument 994 struct isp_device *isp = to_isp_device(prev); preview_config_averager() 1004 * @prev: The preview engine 1015 static void preview_config_input_format(struct isp_prev_device *prev, preview_config_input_format() argument 1018 struct isp_device *isp = to_isp_device(prev); preview_config_input_format() 1030 prev->params.cfa_order = 0; preview_config_input_format() 1033 prev->params.cfa_order = 1; preview_config_input_format() 1036 prev->params.cfa_order = 2; preview_config_input_format() 1039 prev->params.cfa_order = 3; preview_config_input_format() 1052 params = (prev->params.active & OMAP3ISP_PREV_CFA) preview_config_input_format() 1053 ? &prev->params.params[0] : &prev->params.params[1]; preview_config_input_format() 1055 preview_config_cfa(prev, params); preview_config_input_format() 1069 static void preview_config_input_size(struct isp_prev_device *prev, u32 active) preview_config_input_size() argument 1071 const struct v4l2_mbus_framefmt *format = &prev->formats[PREV_PAD_SINK]; preview_config_input_size() 1072 struct isp_device *isp = to_isp_device(prev); preview_config_input_size() 1073 unsigned int sph = prev->crop.left; preview_config_input_size() 1074 unsigned int eph = prev->crop.left + prev->crop.width - 1; preview_config_input_size() 1075 unsigned int slv = prev->crop.top; preview_config_input_size() 1076 unsigned int elv = prev->crop.top + prev->crop.height - 1; preview_config_input_size() 1087 features = (prev->params.params[0].features & active) preview_config_input_size() 1088 | (prev->params.params[1].features & ~active); preview_config_input_size() 1111 * @prev: Preview module 1120 preview_config_inlineoffset(struct isp_prev_device *prev, u32 offset) preview_config_inlineoffset() argument 1122 struct isp_device *isp = to_isp_device(prev); preview_config_inlineoffset() 1134 static void preview_set_inaddr(struct isp_prev_device *prev, u32 addr) preview_set_inaddr() argument 1136 struct isp_device *isp = to_isp_device(prev); preview_set_inaddr() 1147 static void preview_config_outlineoffset(struct isp_prev_device *prev, preview_config_outlineoffset() argument 1150 struct isp_device *isp = to_isp_device(prev); preview_config_outlineoffset() 1162 static void preview_set_outaddr(struct isp_prev_device *prev, u32 addr) preview_set_outaddr() argument 1164 struct isp_device *isp = to_isp_device(prev); preview_set_outaddr() 1169 static void preview_adjust_bandwidth(struct isp_prev_device *prev) preview_adjust_bandwidth() argument 1171 struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); preview_adjust_bandwidth() 1172 struct isp_device *isp = to_isp_device(prev); preview_adjust_bandwidth() 1173 const struct v4l2_mbus_framefmt *ifmt = &prev->formats[PREV_PAD_SINK]; preview_adjust_bandwidth() 1183 if (prev->input != PREVIEW_INPUT_MEMORY) { preview_adjust_bandwidth() 1222 int omap3isp_preview_busy(struct isp_prev_device *prev) omap3isp_preview_busy() argument 1224 struct isp_device *isp = to_isp_device(prev); omap3isp_preview_busy() 1235 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_restore_context() local 1238 prev->params.params[0].update = prev->params.active & update; omap3isp_preview_restore_context() 1239 prev->params.params[1].update = ~prev->params.active & update; omap3isp_preview_restore_context() 1241 preview_setup_hw(prev, update, prev->params.active); omap3isp_preview_restore_context() 1243 prev->params.params[0].update = 0; omap3isp_preview_restore_context() 1244 prev->params.params[1].update = 0; omap3isp_preview_restore_context() 1254 static void preview_print_status(struct isp_prev_device *prev) preview_print_status() argument 1256 struct isp_device *isp = to_isp_device(prev); preview_print_status() 1302 * @prev: pointer to previewer private structure 1304 static void preview_init_params(struct isp_prev_device *prev) preview_init_params() argument 1309 spin_lock_init(&prev->params.lock); preview_init_params() 1311 prev->params.active = ~0; preview_init_params() 1312 prev->params.params[0].busy = 0; preview_init_params() 1313 prev->params.params[0].update = OMAP3ISP_PREV_FEATURES_END - 1; preview_init_params() 1314 prev->params.params[1].busy = 0; preview_init_params() 1315 prev->params.params[1].update = 0; preview_init_params() 1317 params = &prev->params.params[0]; preview_init_params() 1365 * @prev: pointer to previewer private structure 1368 static unsigned int preview_max_out_width(struct isp_prev_device *prev) preview_max_out_width() argument 1370 struct isp_device *isp = to_isp_device(prev); preview_max_out_width() 1385 static void preview_configure(struct isp_prev_device *prev) preview_configure() argument 1387 struct isp_device *isp = to_isp_device(prev); preview_configure() 1394 spin_lock_irqsave(&prev->params.lock, flags); preview_configure() 1396 update = preview_params_lock(prev, 0, false); preview_configure() 1397 active = prev->params.active; preview_configure() 1398 spin_unlock_irqrestore(&prev->params.lock, flags); preview_configure() 1401 format = &prev->formats[PREV_PAD_SINK]; preview_configure() 1404 preview_adjust_bandwidth(prev); preview_configure() 1406 preview_config_input_format(prev, info); preview_configure() 1407 preview_config_input_size(prev, active); preview_configure() 1409 if (prev->input == PREVIEW_INPUT_CCDC) preview_configure() 1410 preview_config_inlineoffset(prev, 0); preview_configure() 1412 preview_config_inlineoffset(prev, ALIGN(format->width, 0x20) * preview_configure() 1415 preview_setup_hw(prev, update, active); preview_configure() 1418 format = &prev->formats[PREV_PAD_SOURCE]; preview_configure() 1420 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_configure() 1427 if (prev->output & PREVIEW_OUTPUT_RESIZER) preview_configure() 1434 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_configure() 1435 preview_config_outlineoffset(prev, preview_configure() 1438 preview_config_averager(prev, 0); preview_configure() 1439 preview_config_ycpos(prev, format->code); preview_configure() 1441 spin_lock_irqsave(&prev->params.lock, flags); preview_configure() 1442 preview_params_unlock(prev, update, false); preview_configure() 1443 spin_unlock_irqrestore(&prev->params.lock, flags); preview_configure() 1450 static void preview_enable_oneshot(struct isp_prev_device *prev) preview_enable_oneshot() argument 1452 struct isp_device *isp = to_isp_device(prev); preview_enable_oneshot() 1458 if (prev->input == PREVIEW_INPUT_MEMORY) preview_enable_oneshot() 1466 void omap3isp_preview_isr_frame_sync(struct isp_prev_device *prev) omap3isp_preview_isr_frame_sync() argument 1474 if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS && omap3isp_preview_isr_frame_sync() 1475 prev->video_out.dmaqueue_flags & ISP_VIDEO_DMAQUEUE_QUEUED) { omap3isp_preview_isr_frame_sync() 1476 preview_enable_oneshot(prev); omap3isp_preview_isr_frame_sync() 1477 isp_video_dmaqueue_flags_clr(&prev->video_out); omap3isp_preview_isr_frame_sync() 1481 static void preview_isr_buffer(struct isp_prev_device *prev) preview_isr_buffer() argument 1483 struct isp_pipeline *pipe = to_isp_pipeline(&prev->subdev.entity); preview_isr_buffer() 1487 if (prev->input == PREVIEW_INPUT_MEMORY) { preview_isr_buffer() 1488 buffer = omap3isp_video_buffer_next(&prev->video_in); preview_isr_buffer() 1490 preview_set_inaddr(prev, buffer->dma); preview_isr_buffer() 1494 if (prev->output & PREVIEW_OUTPUT_MEMORY) { preview_isr_buffer() 1495 buffer = omap3isp_video_buffer_next(&prev->video_out); preview_isr_buffer() 1497 preview_set_outaddr(prev, buffer->dma); preview_isr_buffer() 1503 switch (prev->state) { preview_isr_buffer() 1515 preview_enable_oneshot(prev); preview_isr_buffer() 1529 void omap3isp_preview_isr(struct isp_prev_device *prev) omap3isp_preview_isr() argument 1535 if (omap3isp_module_sync_is_stopping(&prev->wait, &prev->stopping)) omap3isp_preview_isr() 1538 spin_lock_irqsave(&prev->params.lock, flags); omap3isp_preview_isr() 1539 preview_params_switch(prev); omap3isp_preview_isr() 1540 update = preview_params_lock(prev, 0, false); omap3isp_preview_isr() 1541 active = prev->params.active; omap3isp_preview_isr() 1542 spin_unlock_irqrestore(&prev->params.lock, flags); omap3isp_preview_isr() 1544 preview_setup_hw(prev, update, active); omap3isp_preview_isr() 1545 preview_config_input_size(prev, active); omap3isp_preview_isr() 1547 if (prev->input == PREVIEW_INPUT_MEMORY || omap3isp_preview_isr() 1548 prev->output & PREVIEW_OUTPUT_MEMORY) omap3isp_preview_isr() 1549 preview_isr_buffer(prev); omap3isp_preview_isr() 1550 else if (prev->state == ISP_PIPELINE_STREAM_CONTINUOUS) omap3isp_preview_isr() 1551 preview_enable_oneshot(prev); omap3isp_preview_isr() 1553 spin_lock_irqsave(&prev->params.lock, flags); omap3isp_preview_isr() 1554 preview_params_unlock(prev, update, false); omap3isp_preview_isr() 1555 spin_unlock_irqrestore(&prev->params.lock, flags); omap3isp_preview_isr() 1565 struct isp_prev_device *prev = &video->isp->isp_prev; preview_video_queue() local 1568 preview_set_inaddr(prev, buffer->dma); preview_video_queue() 1571 preview_set_outaddr(prev, buffer->dma); preview_video_queue() 1590 struct isp_prev_device *prev = preview_s_ctrl() local 1595 preview_update_brightness(prev, ctrl->val); preview_s_ctrl() 1598 preview_update_contrast(prev, ctrl->val); preview_s_ctrl() 1618 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_ioctl() local 1622 return preview_config(prev, arg); preview_ioctl() 1637 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_stream() local 1638 struct isp_video *video_out = &prev->video_out; preview_set_stream() 1639 struct isp_device *isp = to_isp_device(prev); preview_set_stream() 1640 struct device *dev = to_device(prev); preview_set_stream() 1642 if (prev->state == ISP_PIPELINE_STREAM_STOPPED) { preview_set_stream() 1647 preview_configure(prev); preview_set_stream() 1648 atomic_set(&prev->stopping, 0); preview_set_stream() 1649 preview_print_status(prev); preview_set_stream() 1654 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_set_stream() 1658 !(prev->output & PREVIEW_OUTPUT_MEMORY)) preview_set_stream() 1659 preview_enable_oneshot(prev); preview_set_stream() 1665 if (prev->input == PREVIEW_INPUT_MEMORY) preview_set_stream() 1667 if (prev->output & PREVIEW_OUTPUT_MEMORY) preview_set_stream() 1670 preview_enable_oneshot(prev); preview_set_stream() 1674 if (omap3isp_module_sync_idle(&sd->entity, &prev->wait, preview_set_stream() 1675 &prev->stopping)) preview_set_stream() 1684 prev->state = enable; preview_set_stream() 1689 __preview_get_format(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, __preview_get_format() argument 1693 return v4l2_subdev_get_try_format(&prev->subdev, cfg, pad); __preview_get_format() 1695 return &prev->formats[pad]; __preview_get_format() 1699 __preview_get_crop(struct isp_prev_device *prev, struct v4l2_subdev_pad_config *cfg, __preview_get_crop() argument 1703 return v4l2_subdev_get_try_crop(&prev->subdev, cfg, PREV_PAD_SINK); __preview_get_crop() 1705 return &prev->crop; __preview_get_crop() 1729 * @prev: ISP preview engine 1738 static void preview_try_format(struct isp_prev_device *prev, preview_try_format() argument 1758 if (prev->input == PREVIEW_INPUT_MEMORY) { preview_try_format() 1760 preview_max_out_width(prev)); preview_try_format() 1780 *fmt = *__preview_get_format(prev, cfg, PREV_PAD_SINK, which); preview_try_format() 1798 crop = __preview_get_crop(prev, cfg, which); preview_try_format() 1811 * @prev: ISP preview engine 1821 static void preview_try_crop(struct isp_prev_device *prev, preview_try_crop() argument 1835 if (prev->input == PREVIEW_INPUT_CCDC) { preview_try_crop() 1899 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_enum_frame_size() local 1908 preview_try_format(prev, cfg, fse->pad, &format, fse->which); preview_enum_frame_size() 1918 preview_try_format(prev, cfg, fse->pad, &format, fse->which); preview_enum_frame_size() 1939 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_get_selection() local 1952 format = __preview_get_format(prev, cfg, PREV_PAD_SINK, preview_get_selection() 1954 preview_try_crop(prev, format, &sel->r); preview_get_selection() 1958 sel->r = *__preview_get_crop(prev, cfg, sel->which); preview_get_selection() 1982 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_selection() local 1990 if (prev->state != ISP_PIPELINE_STREAM_STOPPED) preview_set_selection() 1998 sel->r = *__preview_get_crop(prev, cfg, sel->which); preview_set_selection() 2002 format = __preview_get_format(prev, cfg, PREV_PAD_SINK, sel->which); preview_set_selection() 2003 preview_try_crop(prev, format, &sel->r); preview_set_selection() 2004 *__preview_get_crop(prev, cfg, sel->which) = sel->r; preview_set_selection() 2007 format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, sel->which); preview_set_selection() 2008 preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, sel->which); preview_set_selection() 2023 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_get_format() local 2026 format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); preview_get_format() 2044 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_set_format() local 2048 format = __preview_get_format(prev, cfg, fmt->pad, fmt->which); preview_set_format() 2052 preview_try_format(prev, cfg, fmt->pad, &fmt->format, fmt->which); preview_set_format() 2058 crop = __preview_get_crop(prev, cfg, fmt->which); preview_set_format() 2064 preview_try_crop(prev, &fmt->format, crop); preview_set_format() 2067 format = __preview_get_format(prev, cfg, PREV_PAD_SOURCE, preview_set_format() 2069 preview_try_format(prev, cfg, PREV_PAD_SOURCE, format, preview_set_format() 2150 struct isp_prev_device *prev = v4l2_get_subdevdata(sd); preview_link_setup() local 2156 if (prev->input == PREVIEW_INPUT_CCDC) preview_link_setup() 2158 prev->input = PREVIEW_INPUT_MEMORY; preview_link_setup() 2160 if (prev->input == PREVIEW_INPUT_MEMORY) preview_link_setup() 2161 prev->input = PREVIEW_INPUT_NONE; preview_link_setup() 2168 if (prev->input == PREVIEW_INPUT_MEMORY) preview_link_setup() 2170 prev->input = PREVIEW_INPUT_CCDC; preview_link_setup() 2172 if (prev->input == PREVIEW_INPUT_CCDC) preview_link_setup() 2173 prev->input = PREVIEW_INPUT_NONE; preview_link_setup() 2185 if (prev->output & ~PREVIEW_OUTPUT_MEMORY) preview_link_setup() 2187 prev->output |= PREVIEW_OUTPUT_MEMORY; preview_link_setup() 2189 prev->output &= ~PREVIEW_OUTPUT_MEMORY; preview_link_setup() 2196 if (prev->output & ~PREVIEW_OUTPUT_RESIZER) preview_link_setup() 2198 prev->output |= PREVIEW_OUTPUT_RESIZER; preview_link_setup() 2200 prev->output &= ~PREVIEW_OUTPUT_RESIZER; preview_link_setup() 2217 void omap3isp_preview_unregister_entities(struct isp_prev_device *prev) omap3isp_preview_unregister_entities() argument 2219 v4l2_device_unregister_subdev(&prev->subdev); omap3isp_preview_unregister_entities() 2220 omap3isp_video_unregister(&prev->video_in); omap3isp_preview_unregister_entities() 2221 omap3isp_video_unregister(&prev->video_out); omap3isp_preview_unregister_entities() 2224 int omap3isp_preview_register_entities(struct isp_prev_device *prev, omap3isp_preview_register_entities() argument 2230 ret = v4l2_device_register_subdev(vdev, &prev->subdev); omap3isp_preview_register_entities() 2234 ret = omap3isp_video_register(&prev->video_in, vdev); omap3isp_preview_register_entities() 2238 ret = omap3isp_video_register(&prev->video_out, vdev); omap3isp_preview_register_entities() 2245 omap3isp_preview_unregister_entities(prev); omap3isp_preview_register_entities() 2255 * @prev : Pointer to preview structure 2258 static int preview_init_entities(struct isp_prev_device *prev) preview_init_entities() argument 2260 struct v4l2_subdev *sd = &prev->subdev; preview_init_entities() 2261 struct media_pad *pads = prev->pads; preview_init_entities() 2265 prev->input = PREVIEW_INPUT_NONE; preview_init_entities() 2271 v4l2_set_subdevdata(sd, prev); preview_init_entities() 2274 v4l2_ctrl_handler_init(&prev->ctrls, 2); preview_init_entities() 2275 v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_BRIGHTNESS, preview_init_entities() 2278 v4l2_ctrl_new_std(&prev->ctrls, &preview_ctrl_ops, V4L2_CID_CONTRAST, preview_init_entities() 2281 v4l2_ctrl_handler_setup(&prev->ctrls); preview_init_entities() 2282 sd->ctrl_handler = &prev->ctrls; preview_init_entities() 2299 prev->video_in.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; preview_init_entities() 2300 prev->video_in.ops = &preview_video_ops; preview_init_entities() 2301 prev->video_in.isp = to_isp_device(prev); preview_init_entities() 2302 prev->video_in.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; preview_init_entities() 2303 prev->video_in.bpl_alignment = 64; preview_init_entities() 2304 prev->video_out.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; preview_init_entities() 2305 prev->video_out.ops = &preview_video_ops; preview_init_entities() 2306 prev->video_out.isp = to_isp_device(prev); preview_init_entities() 2307 prev->video_out.capture_mem = PAGE_ALIGN(4096 * 4096) * 2 * 3; preview_init_entities() 2308 prev->video_out.bpl_alignment = 32; preview_init_entities() 2310 ret = omap3isp_video_init(&prev->video_in, "preview"); preview_init_entities() 2314 ret = omap3isp_video_init(&prev->video_out, "preview"); preview_init_entities() 2319 ret = media_entity_create_link(&prev->video_in.video.entity, 0, preview_init_entities() 2320 &prev->subdev.entity, PREV_PAD_SINK, 0); preview_init_entities() 2324 ret = media_entity_create_link(&prev->subdev.entity, PREV_PAD_SOURCE, preview_init_entities() 2325 &prev->video_out.video.entity, 0, 0); preview_init_entities() 2332 omap3isp_video_cleanup(&prev->video_out); preview_init_entities() 2334 omap3isp_video_cleanup(&prev->video_in); preview_init_entities() 2336 media_entity_cleanup(&prev->subdev.entity); preview_init_entities() 2347 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_init() local 2349 init_waitqueue_head(&prev->wait); omap3isp_preview_init() 2351 preview_init_params(prev); omap3isp_preview_init() 2353 return preview_init_entities(prev); omap3isp_preview_init() 2358 struct isp_prev_device *prev = &isp->isp_prev; omap3isp_preview_cleanup() local 2360 v4l2_ctrl_handler_free(&prev->ctrls); omap3isp_preview_cleanup() 2361 omap3isp_video_cleanup(&prev->video_in); omap3isp_preview_cleanup() 2362 omap3isp_video_cleanup(&prev->video_out); omap3isp_preview_cleanup() 2363 media_entity_cleanup(&prev->subdev.entity); omap3isp_preview_cleanup()
|
/linux-4.1.27/arch/avr32/include/asm/ |
H A D | switch_to.h | 18 #define finish_arch_switch(prev) \ 20 ocd_write(PID, prev->pid); \ 26 * switch_to(prev, next, last) should switch from task `prev' to task 27 * `next'. `prev' will never be the same as `next'. 39 #define switch_to(prev, next, last) \ 41 last = __switch_to(prev, &prev->thread.cpu_context + 1, \
|
H A D | mmu_context.h | 108 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument 112 if (likely(prev != next)) { switch_mm() 122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.1.27/arch/frv/include/asm/ |
H A D | switch_to.h | 18 * switch_to(prev, next) should switch from task `prev' to `next' 19 * `prev' will never be the same as `next'. 25 struct task_struct *prev); 27 #define switch_to(prev, next, last) \ 29 (prev)->thread.sched_lr = \ 31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
H A D | mmu_context.h | 35 #define switch_mm(prev, next, tsk) \ 37 if (prev != next) \ 38 change_mm_context(&prev->context, &next->context, next->pgd); \ 41 #define activate_mm(prev, next) \ 43 change_mm_context(&prev->context, &next->context, next->pgd); \
|
/linux-4.1.27/arch/mips/include/asm/ |
H A D | switch_to.h | 31 * @prev: The task previously executed. 34 * @fp_save: Which, if any, FP context to save for prev. 36 * This function is used whilst scheduling to save the context of prev & load 37 * the context of next. Returns prev. 39 extern asmlinkage struct task_struct *resume(struct task_struct *prev, 60 #define __mips_mt_fpaff_switch_to(prev) \ 62 struct thread_info *__prev_ti = task_thread_info(prev); \ 66 (!(KSTK_STATUS(prev) & ST0_CU1))) { \ 68 prev->cpus_allowed = prev->thread.user_cpus_allowed; \ 74 #define __mips_mt_fpaff_switch_to(prev) do { (void) (prev); } while (0) 86 #define switch_to(prev, next, last) \ 90 __mips_mt_fpaff_switch_to(prev); \ 92 __save_dsp(prev); \ 93 if (cop2_present && (KSTK_STATUS(prev) & ST0_CU2)) { \ 95 KSTK_STATUS(prev) &= ~ST0_CU2; \ 98 cop2_save(prev); \ 102 if (test_and_clear_tsk_thread_flag(prev, TIF_USEDFPU)) \ 104 if (test_and_clear_tsk_thread_flag(prev, TIF_USEDMSA)) \ 106 (last) = resume(prev, next, task_thread_info(next), __fpsave); \ 109 #define finish_arch_switch(prev) \
|
H A D | mmu_context.h | 139 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 157 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm() 179 activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 194 cpumask_clear_cpu(cpu, mm_cpumask(prev)); activate_mm()
|
/linux-4.1.27/arch/cris/include/asm/ |
H A D | switch_to.h | 8 extern struct task_struct *resume(struct task_struct *prev, struct task_struct *next, int); 9 #define switch_to(prev,next,last) last = resume(prev,next, \
|
H A D | mmu_context.h | 9 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 14 #define activate_mm(prev,next) switch_mm((prev),(next),NULL)
|
H A D | fasttimer.h | 21 struct fast_timer *prev; member in struct:fast_timer
|
/linux-4.1.27/arch/mn10300/include/asm/ |
H A D | switch_to.h | 22 #define switch_fpu(prev, next) \ 24 if ((prev)->thread.fpu_flags & THREAD_HAS_FPU) { \ 25 (prev)->thread.fpu_flags &= ~THREAD_HAS_FPU; \ 26 (prev)->thread.uregs->epsw &= ~EPSW_FE; \ 27 fpu_save(&(prev)->thread.fpu_state); \ 31 #define switch_fpu(prev, next) do {} while (0) 36 struct task_struct *__switch_to(struct thread_struct *prev, 40 #define switch_to(prev, next, last) \ 42 switch_fpu(prev, next); \ 44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|
H A D | mmu_context.h | 139 * @prev: The outgoing MM context. 143 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 148 if (prev != next) { switch_mm() 159 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.1.27/kernel/locking/ |
H A D | osq_lock.c | 38 struct optimistic_spin_node *prev) osq_wait_next() 45 * If there is a prev node in queue, then the 'old' value will be osq_wait_next() 46 * the prev node's CPU #, else it's set to OSQ_UNLOCKED_VAL since if osq_wait_next() 49 old = prev ? prev->cpu : OSQ_UNLOCKED_VAL; osq_wait_next() 55 * We were the last queued, we moved @lock back. @prev osq_wait_next() 65 * @node->next might complete Step-A and think its @prev is osq_wait_next() 87 struct optimistic_spin_node *prev, *next; osq_lock() local 99 prev = decode_cpu(old); osq_lock() 100 node->prev = prev; osq_lock() 101 WRITE_ONCE(prev->next, node); osq_lock() 104 * Normally @prev is untouchable after the above store; because at that osq_lock() 125 * Step - A -- stabilize @prev osq_lock() 127 * Undo our @prev->next assignment; this will make @prev's osq_lock() 133 if (prev->next == node && osq_lock() 134 cmpxchg(&prev->next, node, NULL) == node) osq_lock() 149 * case its step-C will write us a new @node->prev pointer. osq_lock() 151 prev = READ_ONCE(node->prev); osq_lock() 158 * back to @prev. osq_lock() 161 next = osq_wait_next(lock, node, prev); osq_lock() 168 * @prev is stable because its still waiting for a new @prev->next osq_lock() 173 WRITE_ONCE(next->prev, prev); osq_lock() 174 WRITE_ONCE(prev->next, next); osq_lock() 36 osq_wait_next(struct optimistic_spin_queue *lock, struct optimistic_spin_node *node, struct optimistic_spin_node *prev) osq_wait_next() argument
|
H A D | mcs_spinlock.h | 63 struct mcs_spinlock *prev; mcs_spin_lock() local 69 prev = xchg(lock, node); mcs_spin_lock() 70 if (likely(prev == NULL)) { mcs_spin_lock() 81 WRITE_ONCE(prev->next, node); mcs_spin_lock()
|
H A D | mutex.h | 17 __list_del((waiter)->list.prev, (waiter)->list.next)
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | mmu_context.h | 72 static inline void switch_mmu(struct mm_struct *prev, struct mm_struct *next) switch_mmu() argument 79 /* prev->context == prev->pgd in the case where we are initially switch_mmu() 81 if (prev->context.pgd_base != (unsigned long) prev->pgd) { switch_mmu() 83 ((pgd_t *) prev->context.pgd_base)[i] = prev->pgd[i]; switch_mmu() 85 prev->pgd = (pgd_t *)mmu_get_base(); switch_mmu() 87 next->pgd = prev->pgd; switch_mmu() 88 prev->pgd = (pgd_t *) prev->context.pgd_base; switch_mmu() 98 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 101 if (prev != next) switch_mm() 102 switch_mmu(prev, next); switch_mm()
|
/linux-4.1.27/drivers/fmc/ |
H A D | fmc-dump.c | 25 const uint8_t *prev) dump_line() 29 if (!prev || memcmp(line, prev, LINELEN)) { dump_line() 39 if (line == prev + LINELEN) dump_line() 41 return prev; dump_line() 46 const uint8_t *line, *prev; fmc_dump_eeprom() local 58 prev = NULL; fmc_dump_eeprom() 60 prev = dump_line(i, line, prev); fmc_dump_eeprom() 65 const uint8_t *line, *prev; fmc_dump_sdb() local 96 prev = NULL; fmc_dump_sdb() 98 prev = dump_line(i, line, prev); fmc_dump_sdb() 24 dump_line(int addr, const uint8_t *line, const uint8_t *prev) dump_line() argument
|
/linux-4.1.27/arch/powerpc/include/asm/ |
H A D | cmpxchg.h | 18 unsigned long prev; __xchg_u32() local 27 : "=&r" (prev), "+m" (*(volatile unsigned int *)p) __xchg_u32() 31 return prev; __xchg_u32() 43 unsigned long prev; __xchg_u32_local() local 50 : "=&r" (prev), "+m" (*(volatile unsigned int *)p) __xchg_u32_local() 54 return prev; __xchg_u32_local() 61 unsigned long prev; __xchg_u64() local 70 : "=&r" (prev), "+m" (*(volatile unsigned long *)p) __xchg_u64() 74 return prev; __xchg_u64() 80 unsigned long prev; __xchg_u64_local() local 87 : "=&r" (prev), "+m" (*(volatile unsigned long *)p) __xchg_u64_local() 91 return prev; __xchg_u64_local() 152 unsigned int prev; __cmpxchg_u32() local 165 : "=&r" (prev), "+m" (*p) __cmpxchg_u32() 169 return prev; __cmpxchg_u32() 176 unsigned int prev; __cmpxchg_u32_local() local 187 : "=&r" (prev), "+m" (*p) __cmpxchg_u32_local() 191 return prev; __cmpxchg_u32_local() 198 unsigned long prev; __cmpxchg_u64() local 210 : "=&r" (prev), "+m" (*p) __cmpxchg_u64() 214 return prev; __cmpxchg_u64() 221 unsigned long prev; __cmpxchg_u64_local() local 231 : "=&r" (prev), "+m" (*p) __cmpxchg_u64_local() 235 return prev; __cmpxchg_u64_local()
|
H A D | switch_to.h | 13 #define switch_to(prev, next, last) ((last) = __switch_to((prev), (next))) 16 extern struct task_struct *_switch(struct thread_struct *prev, 19 static inline void save_early_sprs(struct thread_struct *prev) save_early_sprs() argument 22 prev->tar = mfspr(SPRN_TAR); save_early_sprs() 24 prev->dscr = mfspr(SPRN_DSCR); save_early_sprs() 27 static inline void save_early_sprs(struct thread_struct *prev) {} save_early_sprs() argument
|
H A D | mmu_context.h | 20 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next); 42 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 58 if (prev == next) switch_mm() 62 /* Switch coprocessor context only if prev or next uses a coprocessor */ switch_mm() 63 if (prev->context.acop || next->context.acop) switch_mm() 82 switch_mmu_context(prev, next); switch_mm() 93 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 98 switch_mm(prev, next, current); activate_mm()
|
H A D | futex.h | 90 u32 prev; futex_atomic_cmpxchg_inatomic() local 112 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic() 116 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.1.27/arch/um/os-Linux/ |
H A D | irq.c | 82 struct irq_fd **prev; os_free_irq_by_cb() local 85 prev = &active_fds; os_free_irq_by_cb() 86 while (*prev != NULL) { os_free_irq_by_cb() 87 if ((*test)(*prev, arg)) { os_free_irq_by_cb() 88 struct irq_fd *old_fd = *prev; os_free_irq_by_cb() 90 (pollfds[i].fd != (*prev)->fd)) { os_free_irq_by_cb() 94 (*prev)->fd, pollfds[i].fd); os_free_irq_by_cb() 107 *last_irq_ptr2 = prev; os_free_irq_by_cb() 109 *prev = (*prev)->next; os_free_irq_by_cb() 115 prev = &(*prev)->next; os_free_irq_by_cb()
|
/linux-4.1.27/arch/microblaze/include/asm/ |
H A D | switch_to.h | 15 extern struct task_struct *_switch_to(struct thread_info *prev, 18 #define switch_to(prev, next, last) \ 20 (last) = _switch_to(task_thread_info(prev), \
|
H A D | futex.h | 101 u32 prev; futex_atomic_cmpxchg_inatomic() local 120 : "+r" (ret), "=&r" (prev), "=&r"(cmp) \ futex_atomic_cmpxchg_inatomic() 123 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.1.27/arch/ia64/include/asm/ |
H A D | switch_to.h | 44 #define __switch_to(prev,next,last) do { \ 45 if (IA64_HAS_EXTRA_STATE(prev)) \ 46 ia64_save_extra(prev); \ 60 # define switch_to(prev,next,last) do { \ 61 if (ia64_psr(task_pt_regs(prev))->mfh && ia64_is_local_fpu_owner(prev)) { \ 62 ia64_psr(task_pt_regs(prev))->mfh = 0; \ 63 (prev)->thread.flags |= IA64_THREAD_FPH_VALID; \ 64 __ia64_save_fpu((prev)->thread.fph); \ 66 __switch_to(prev, next, last); \ 76 # define switch_to(prev,next,last) __switch_to(prev, next, last)
|
H A D | futex.h | 110 unsigned long prev; futex_atomic_cmpxchg_inatomic() local 117 : "+r" (r8), "=&r" (prev) futex_atomic_cmpxchg_inatomic() 121 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.1.27/include/linux/ |
H A D | user-return-notifier.h | 18 static inline void propagate_user_return_notify(struct task_struct *prev, propagate_user_return_notify() argument 21 if (test_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY)) { propagate_user_return_notify() 22 clear_tsk_thread_flag(prev, TIF_USER_RETURN_NOTIFY); propagate_user_return_notify() 38 static inline void propagate_user_return_notify(struct task_struct *prev, propagate_user_return_notify() argument
|
H A D | vtime.h | 42 extern void vtime_task_switch(struct task_struct *prev); vtime_accounting_enabled() 44 extern void vtime_common_task_switch(struct task_struct *prev); vtime_task_switch() 45 static inline void vtime_task_switch(struct task_struct *prev) vtime_task_switch() argument 48 vtime_common_task_switch(prev); vtime_task_switch() 69 static inline void vtime_task_switch(struct task_struct *prev) { } vtime_account_system() argument
|
H A D | context_tracking.h | 17 extern void __context_tracking_task_switch(struct task_struct *prev, 54 static inline void context_tracking_task_switch(struct task_struct *prev, context_tracking_task_switch() argument 58 __context_tracking_task_switch(prev, next); context_tracking_task_switch() 65 static inline void context_tracking_task_switch(struct task_struct *prev, context_tracking_task_switch() argument
|
H A D | list.h | 15 * sometimes we already know the next/prev entries and we can 28 list->prev = list; INIT_LIST_HEAD() 35 * the prev/next entries already! 39 struct list_head *prev, __list_add() 42 next->prev = new; __list_add() 44 new->prev = prev; __list_add() 45 prev->next = new; __list_add() 49 struct list_head *prev, 77 __list_add(new, head->prev, head); list_add_tail() 81 * Delete a list entry by making the prev/next entries 85 * the prev/next entries already! 87 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument 89 next->prev = prev; __list_del() 90 prev->next = next; __list_del() 102 __list_del(entry->prev, entry->next); __list_del_entry() 107 __list_del(entry->prev, entry->next); list_del() 109 entry->prev = LIST_POISON2; list_del() 127 new->next->prev = new; list_replace() 128 new->prev = old->prev; list_replace() 129 new->prev->next = new; list_replace() 198 * in the process of modifying either member (next or prev) 208 return (next == head) && (next == head->prev); list_empty_careful() 231 return !list_empty(head) && (head->next == head->prev); list_is_singular() 239 list->next->prev = list; __list_cut_position() 240 list->prev = entry; __list_cut_position() 243 new_first->prev = head; __list_cut_position() 275 struct list_head *prev, __list_splice() 279 struct list_head *last = list->prev; __list_splice() 281 first->prev = prev; __list_splice() 282 prev->next = first; __list_splice() 285 next->prev = last; __list_splice() 309 __list_splice(list, head->prev, head); list_splice_tail() 340 __list_splice(list, head->prev, head); list_splice_tail_init() 374 list_entry((ptr)->prev, type, member) 396 * list_prev_entry - get the prev element in list 401 list_entry((pos)->member.prev, typeof(*(pos)), member) 417 for (pos = (head)->prev; pos != (head); pos = pos->prev) 436 for (pos = (head)->prev, n = pos->prev; \ 438 pos = n, n = pos->prev) 659 struct hlist_node *prev) hlist_add_behind() 661 n->next = prev->next; hlist_add_behind() 662 prev->next = n; hlist_add_behind() 663 n->pprev = &prev->next; hlist_add_behind() 38 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 274 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument 658 hlist_add_behind(struct hlist_node *n, struct hlist_node *prev) hlist_add_behind() argument
|
H A D | rculist.h | 33 ACCESS_ONCE(list->prev) = list; INIT_LIST_HEAD_RCU() 46 * the prev/next entries already! 50 struct list_head *prev, struct list_head *next) __list_add_rcu() 53 new->prev = prev; __list_add_rcu() 54 rcu_assign_pointer(list_next_rcu(prev), new); __list_add_rcu() 55 next->prev = new; __list_add_rcu() 59 struct list_head *prev, struct list_head *next); 102 __list_add_rcu(new, head->prev, head); list_add_tail_rcu() 132 entry->prev = LIST_POISON2; list_del_rcu() 175 new->prev = old->prev; list_replace_rcu() 176 rcu_assign_pointer(list_next_rcu(new->prev), new); list_replace_rcu() 177 new->next->prev = new; list_replace_rcu() 178 old->prev = LIST_POISON2; list_replace_rcu() 203 struct list_head *last = list->prev; list_splice_init_rcu() 230 * permitted to traverse the prev pointers without excluding list_splice_init_rcu() 236 first->prev = head; list_splice_init_rcu() 237 at->prev = last; list_splice_init_rcu() 437 * @prev: the existing element to add the new element after. 453 struct hlist_node *prev) hlist_add_behind_rcu() 455 n->next = prev->next; hlist_add_behind_rcu() 456 n->pprev = &prev->next; hlist_add_behind_rcu() 457 rcu_assign_pointer(hlist_next_rcu(prev), n); hlist_add_behind_rcu() 49 __list_add_rcu(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add_rcu() argument 452 hlist_add_behind_rcu(struct hlist_node *n, struct hlist_node *prev) hlist_add_behind_rcu() argument
|
H A D | osq_lock.h | 9 struct optimistic_spin_node *next, *prev; member in struct:optimistic_spin_node
|
H A D | of_pdt.h | 20 * If prev is NULL, the first property will be returned. 22 int (*nextprop)(phandle node, char *prev, char *buf);
|
H A D | agpgart.h | 93 struct agp_client *prev; member in struct:agp_client 101 struct agp_controller *prev; member in struct:agp_controller 116 struct agp_file_private *prev; member in struct:agp_file_private
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | switch_to_64.h | 19 struct task_struct *sh64_switch_to(struct task_struct *prev, 24 #define switch_to(prev,next,last) \ 30 last = sh64_switch_to(prev, &prev->thread, next, \
|
H A D | switch_to_32.h | 63 struct task_struct *__switch_to(struct task_struct *prev, 69 #define switch_to(prev, next, last) \ 79 if (is_dsp_enabled(prev)) \ 80 __save_dsp(prev); \ 82 __ts1 = (u32 *)&prev->thread.sp; \ 83 __ts2 = (u32 *)&prev->thread.pc; \ 84 __ts4 = (u32 *)prev; \ 128 #define finish_arch_switch(prev) \ 130 if (is_dsp_enabled(prev)) \ 131 __restore_dsp(prev); \
|
H A D | futex-irq.h | 96 u32 prev = 0; atomic_futex_op_cmpxchg_inatomic() local 100 ret = get_user(prev, uaddr); atomic_futex_op_cmpxchg_inatomic() 101 if (!ret && oldval == prev) atomic_futex_op_cmpxchg_inatomic() 106 *uval = prev; atomic_futex_op_cmpxchg_inatomic()
|
H A D | mmu_context.h | 118 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument 124 if (likely(prev != next)) { switch_mm() 133 #define activate_mm(prev, next) switch_mm((prev),(next),NULL)
|
/linux-4.1.27/arch/c6x/include/asm/ |
H A D | switch_to.h | 20 asmlinkage void *__switch_to(struct thread_struct *prev, 24 #define switch_to(prev, next, last) \ 27 (last) = __switch_to(&(prev)->thread, \ 28 &(next)->thread, (prev)); \
|
/linux-4.1.27/arch/m32r/include/asm/ |
H A D | switch_to.h | 13 * switch_to(prev, next) should switch from task `prev' to `next' 14 * `prev' will never be the same as `next'. 16 * `next' and `prev' should be struct task_struct, but it isn't always defined 28 #define switch_to(prev, next, last) do { \ 37 " push %1 ; store `prev' on new stack \n" \ 44 : "0" (prev), \ 45 "r" (&(prev->thread.sp)), "r" (&(next->thread.sp)), \ 46 "r" (&(prev->thread.lr)), "r" (&(next->thread.lr)) \
|
H A D | mmu_context.h | 121 static inline void switch_mm(struct mm_struct *prev, switch_mm() argument 128 if (prev != next) { switch_mm() 145 #define activate_mm(prev, next) \ 146 switch_mm((prev), (next), NULL) 155 #define switch_mm(prev,next,tsk) do { } while (0) 157 #define activate_mm(prev,next) do { } while (0)
|
/linux-4.1.27/tools/usb/usbip/libsrc/ |
H A D | list.h | 13 * sometimes we already know the next/prev entries and we can 19 struct list_head *next, *prev; member in struct:list_head 30 list->prev = list; INIT_LIST_HEAD() 37 * the prev/next entries already! 40 struct list_head *prev, __list_add() 43 next->prev = new; __list_add() 45 new->prev = prev; __list_add() 46 prev->next = new; __list_add() 63 * Delete a list entry by making the prev/next entries 67 * the prev/next entries already! 69 static inline void __list_del(struct list_head * prev, struct list_head * next) __list_del() argument 71 next->prev = prev; __list_del() 72 prev->next = next; __list_del() 87 __list_del(entry->prev, entry->next); __list_del_entry() 92 __list_del(entry->prev, entry->next); list_del() 94 entry->prev = LIST_POISON2; list_del() 39 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.1.27/arch/arc/include/asm/ |
H A D | switch_to.h | 31 #define switch_to(prev, next, last) \ 33 ARC_FPU_PREV(prev, next); \ 34 last = __switch_to(prev, next);\
|
H A D | cmpxchg.h | 22 unsigned long prev; __cmpxchg() local 36 : "=&r"(prev) /* Early clobber, to prevent reg reuse */ __cmpxchg() 44 return prev; __cmpxchg() 53 int prev; __cmpxchg() local 60 prev = *p; __cmpxchg() 61 if (prev == expected) __cmpxchg() 64 return prev; __cmpxchg()
|
/linux-4.1.27/scripts/kconfig/ |
H A D | list.h | 24 struct list_head *next, *prev; member in struct:list_head 79 * the prev/next entries already! 82 struct list_head *prev, __list_add() 85 next->prev = _new; __list_add() 87 _new->prev = prev; __list_add() 88 prev->next = _new; __list_add() 101 __list_add(_new, head->prev, head); list_add_tail() 105 * Delete a list entry by making the prev/next entries 109 * the prev/next entries already! 111 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 113 next->prev = prev; __list_del() 114 prev->next = next; __list_del() 127 __list_del(entry->prev, entry->next); list_del() 129 entry->prev = (struct list_head*)LIST_POISON2; list_del() 81 __list_add(struct list_head *_new, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.1.27/arch/xtensa/include/asm/ |
H A D | nommu_context.h | 18 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 22 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument
|
H A D | switch_to.h | 17 #define switch_to(prev,next,last) \ 19 (last) = _switch_to(prev, next); \
|
H A D | futex.h | 112 u32 prev; futex_atomic_cmpxchg_inatomic() local 138 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic() 142 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
H A D | mmu_context.h | 123 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 133 if (migrated || prev != next) switch_mm() 137 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.1.27/arch/tile/include/asm/ |
H A D | switch_to.h | 40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next))) 41 extern struct task_struct *_switch_to(struct task_struct *prev, 45 extern struct task_struct *__switch_to(struct task_struct *prev, 57 #define finish_arch_switch(prev) do { \ 58 if (unlikely((prev)->state == TASK_DEAD)) \ 60 ((prev)->pid << _SIM_CONTROL_OPERATOR_BITS)); \ 75 struct task_struct *sim_notify_fork(struct task_struct *prev);
|
H A D | mmu_context.h | 91 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 94 if (likely(prev != next)) { switch_mm() 107 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm() 114 check_mm_caching(prev, next); switch_mm()
|
/linux-4.1.27/arch/x86/include/asm/ |
H A D | cmpxchg_32.h | 27 u64 prev = *ptr; set_64bit() local 32 : "=m" (*ptr), "+A" (prev) set_64bit() 48 u64 prev; __cmpxchg64() local 50 : "=A" (prev), __cmpxchg64() 56 return prev; __cmpxchg64() 61 u64 prev; __cmpxchg64_local() local 63 : "=A" (prev), __cmpxchg64_local() 69 return prev; __cmpxchg64_local()
|
H A D | switch_to.h | 5 __visible struct task_struct *__switch_to(struct task_struct *prev, 31 #define switch_to(prev, next, last) \ 55 : [prev_sp] "=m" (prev->thread.sp), \ 56 [prev_ip] "=m" (prev->thread.ip), \ 70 [prev] "a" (prev), \ 108 #define switch_to(prev, next, last) \ 110 "movq %%rsp,%P[threadrsp](%[prev])\n\t" /* save RSP */ \ 122 : [next] "S" (next), [prev] "D" (prev), \
|
H A D | mmu_context.h | 15 static inline void paravirt_activate_mm(struct mm_struct *prev, paravirt_activate_mm() argument 95 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 100 if (likely(prev != next)) { switch_mm() 140 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm() 148 * It's possible that prev->context.ldt doesn't match switch_mm() 149 * the LDT register. This can happen if leave_mm(prev) switch_mm() 151 * prev->context.ldt but suppressed an IPI to this CPU. switch_mm() 152 * In this case, prev->context.ldt != NULL, because we switch_mm() 155 * prev->context.ldt, because mms never share an LDT. switch_mm() 157 if (unlikely(prev->context.ldt != next->context.ldt)) switch_mm() 191 #define activate_mm(prev, next) \ 193 paravirt_activate_mm((prev), (next)); \ 194 switch_mm((prev), (next), NULL); \
|
/linux-4.1.27/arch/parisc/lib/ |
H A D | bitops.c | 62 unsigned long prev; __cmpxchg_u64() local 65 if ((prev = *ptr) == old) __cmpxchg_u64() 68 return prev; __cmpxchg_u64() 75 unsigned int prev; __cmpxchg_u32() local 78 if ((prev = *ptr) == old) __cmpxchg_u32() 81 return (unsigned long)prev; __cmpxchg_u32()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | mm.c | 53 struct nvkm_mm_node *prev = node(this, prev); nvkm_mm_free() local 56 if (prev && prev->type == NVKM_MM_TYPE_NONE) { nvkm_mm_free() 57 prev->length += this->length; nvkm_mm_free() 59 kfree(this); this = prev; nvkm_mm_free() 72 list_for_each_entry(prev, &mm->free, fl_entry) { nvkm_mm_free() 73 if (this->offset < prev->offset) nvkm_mm_free() 77 list_add_tail(&this->fl_entry, &prev->fl_entry); nvkm_mm_free() 114 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_head() local 129 prev = node(this, prev); nvkm_mm_head() 130 if (prev && prev->type != type) nvkm_mm_head() 188 struct nvkm_mm_node *prev, *this, *next; nvkm_mm_tail() local 202 prev = node(this, prev); nvkm_mm_tail() 203 if (prev && prev->type != type) nvkm_mm_tail() 240 struct nvkm_mm_node *node, *prev; nvkm_mm_init() local 244 prev = list_last_entry(&mm->nodes, typeof(*node), nl_entry); nvkm_mm_init() 245 next = prev->offset + prev->length; nvkm_mm_init()
|
/linux-4.1.27/drivers/net/fddi/skfp/ |
H A D | smttimer.c | 41 struct smt_timer **prev ; smt_timer_stop() local 51 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { smt_timer_stop() 53 *prev = tm->tm_next ; smt_timer_stop() 65 struct smt_timer **prev ; smt_timer_start() local 92 for (prev = &smc->t.st_queue ; (tm = *prev) ; prev = &tm->tm_next ) { smt_timer_start() 99 *prev = timer ; smt_timer_start()
|
/linux-4.1.27/net/bridge/ |
H A D | br_forward.c | 24 static int deliver_clone(const struct net_bridge_port *prev, 134 static int deliver_clone(const struct net_bridge_port *prev, deliver_clone() argument 147 __packet_hook(prev, skb); deliver_clone() 152 struct net_bridge_port *prev, struct net_bridge_port *p, maybe_deliver() 160 return prev; maybe_deliver() 162 if (!prev) maybe_deliver() 165 err = deliver_clone(prev, skb, __packet_hook); maybe_deliver() 181 struct net_bridge_port *prev; br_flood() local 183 prev = NULL; br_flood() 197 prev = maybe_deliver(prev, p, skb, __packet_hook); br_flood() 198 if (IS_ERR(prev)) br_flood() 202 if (!prev) br_flood() 206 deliver_clone(prev, skb, __packet_hook); br_flood() 208 __packet_hook(prev, skb); br_flood() 240 struct net_bridge_port *prev = NULL; br_multicast_flood() local 256 prev = maybe_deliver(prev, port, skb, __packet_hook); br_multicast_flood() 257 if (IS_ERR(prev)) br_multicast_flood() 266 if (!prev) br_multicast_flood() 270 deliver_clone(prev, skb, __packet_hook); br_multicast_flood() 272 __packet_hook(prev, skb); br_multicast_flood() 151 maybe_deliver( struct net_bridge_port *prev, struct net_bridge_port *p, struct sk_buff *skb, void (*__packet_hook)(const struct net_bridge_port *p, struct sk_buff *skb)) maybe_deliver() argument
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | mkregtable.c | 36 * sometimes we already know the next/prev entries and we can 42 struct list_head *next, *prev; member in struct:list_head 53 list->prev = list; INIT_LIST_HEAD() 60 * the prev/next entries already! 64 struct list_head *prev, struct list_head *next) __list_add() 66 next->prev = new; __list_add() 68 new->prev = prev; __list_add() 69 prev->next = new; __list_add() 73 struct list_head *prev, struct list_head *next); 99 __list_add(new, head->prev, head); list_add_tail() 103 * Delete a list entry by making the prev/next entries 107 * the prev/next entries already! 109 static inline void __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 111 next->prev = prev; __list_del() 112 prev->next = next; __list_del() 124 __list_del(entry->prev, entry->next); list_del() 126 entry->prev = (void *)0xBEEFDEAD; list_del() 142 new->next->prev = new; list_replace() 143 new->prev = old->prev; list_replace() 144 new->prev->next = new; list_replace() 160 __list_del(entry->prev, entry->next); list_del_init() 171 __list_del(list->prev, list->next); list_move() 183 __list_del(list->prev, list->next); list_move_tail() 213 * in the process of modifying either member (next or prev) 223 return (next == head) && (next == head->prev); list_empty_careful() 232 return !list_empty(head) && (head->next == head->prev); list_is_singular() 241 list->next->prev = list; __list_cut_position() 242 list->prev = entry; __list_cut_position() 245 new_first->prev = head; __list_cut_position() 277 struct list_head *prev, struct list_head *next) __list_splice() 280 struct list_head *last = list->prev; __list_splice() 282 first->prev = prev; __list_splice() 283 prev->next = first; __list_splice() 286 next->prev = last; __list_splice() 310 __list_splice(list, head->prev, head); list_splice_tail() 341 __list_splice(list, head->prev, head); list_splice_tail_init() 381 for (pos = (head)->prev; prefetch(pos->prev), pos != (head); \ 382 pos = pos->prev) 401 for (pos = (head)->prev, n = pos->prev; \ 402 prefetch(pos->prev), pos != (head); \ 403 pos = n, n = pos->prev) 423 for (pos = list_entry((head)->prev, typeof(*pos), member); \ 424 prefetch(pos->member.prev), &pos->member != (head); \ 425 pos = list_entry(pos->member.prev, typeof(*pos), member)) 462 for (pos = list_entry(pos->member.prev, typeof(*pos), member); \ 463 prefetch(pos->member.prev), &pos->member != (head); \ 464 pos = list_entry(pos->member.prev, typeof(*pos), member)) 533 for (pos = list_entry((head)->prev, typeof(*pos), member), \ 534 n = list_entry(pos->member.prev, typeof(*pos), member); \ 536 pos = n, n = list_entry(n->member.prev, typeof(*n), member)) 63 __list_add(struct list_head *new, struct list_head *prev, struct list_head *next) __list_add() argument 276 __list_splice(const struct list_head *list, struct list_head *prev, struct list_head *next) __list_splice() argument
|
H A D | radeon_mem.c | 55 newblock->prev = p; split_block() 56 p->next->prev = newblock; split_block() 72 newblock->prev = p; split_block() 73 p->next->prev = newblock; split_block() 121 p->next->prev = p; free_block() 125 if (p->prev->file_priv == NULL) { free_block() 126 struct mem_block *q = p->prev; free_block() 129 q->next->prev = q; free_block() 152 blocks->next = blocks->prev = *heap; init_heap() 155 (*heap)->next = (*heap)->prev = blocks; init_heap() 181 p->next->prev = p; list_for_each()
|
/linux-4.1.27/arch/cris/arch-v32/mm/ |
H A D | intmem.c | 107 struct intmem_allocation *prev = crisv32_intmem_free() local 108 list_entry(allocation->entry.prev, crisv32_intmem_free() 115 /* Join with prev and/or next if also free */ crisv32_intmem_free() 116 if ((prev != &intmem_allocations) && crisv32_intmem_free() 117 (prev->status == STATUS_FREE)) { crisv32_intmem_free() 118 prev->size += allocation->size; crisv32_intmem_free() 121 allocation = prev; crisv32_intmem_free()
|
/linux-4.1.27/mm/ |
H A D | interval_tree.c | 28 /* Insert node immediately after prev in the interval tree */ vma_interval_tree_insert_after() 30 struct vm_area_struct *prev, vma_interval_tree_insert_after() 37 VM_BUG_ON_VMA(vma_start_pgoff(node) != vma_start_pgoff(prev), node); vma_interval_tree_insert_after() 39 if (!prev->shared.rb.rb_right) { vma_interval_tree_insert_after() 40 parent = prev; vma_interval_tree_insert_after() 41 link = &prev->shared.rb.rb_right; vma_interval_tree_insert_after() 43 parent = rb_entry(prev->shared.rb.rb_right, vma_interval_tree_insert_after() 29 vma_interval_tree_insert_after(struct vm_area_struct *node, struct vm_area_struct *prev, struct rb_root *root) vma_interval_tree_insert_after() argument
|
H A D | madvise.c | 46 struct vm_area_struct **prev, madvise_behavior() 99 *prev = vma; madvise_behavior() 104 *prev = vma_merge(mm, *prev, start, end, new_flags, vma->anon_vma, madvise_behavior() 106 if (*prev) { madvise_behavior() 107 vma = *prev; madvise_behavior() 111 *prev = vma; madvise_behavior() 219 struct vm_area_struct **prev, madvise_willneed() 226 *prev = vma; madvise_willneed() 232 *prev = vma; madvise_willneed() 247 *prev = vma; madvise_willneed() 277 struct vm_area_struct **prev, madvise_dontneed() 280 *prev = vma; madvise_dontneed() 293 struct vm_area_struct **prev, madvise_remove() 300 *prev = NULL; /* tell sys_madvise we drop mmap_sem */ madvise_remove() 372 madvise_vma(struct vm_area_struct *vma, struct vm_area_struct **prev, madvise_vma() argument 377 return madvise_remove(vma, prev, start, end); madvise_vma() 379 return madvise_willneed(vma, prev, start, end); madvise_vma() 381 return madvise_dontneed(vma, prev, start, end); madvise_vma() 383 return madvise_behavior(vma, prev, start, end, behavior); madvise_vma() 461 struct vm_area_struct *vma, *prev; SYSCALL_DEFINE3() local 502 vma = find_vma_prev(current->mm, start, &prev); SYSCALL_DEFINE3() 504 prev = vma; SYSCALL_DEFINE3() 527 error = madvise_vma(vma, &prev, start, tmp, behavior); SYSCALL_DEFINE3() 531 if (prev && start < prev->vm_end) SYSCALL_DEFINE3() 532 start = prev->vm_end; SYSCALL_DEFINE3() 536 if (prev) SYSCALL_DEFINE3() 537 vma = prev->vm_next; SYSCALL_DEFINE3() 45 madvise_behavior(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end, int behavior) madvise_behavior() argument 218 madvise_willneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_willneed() argument 276 madvise_dontneed(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_dontneed() argument 292 madvise_remove(struct vm_area_struct *vma, struct vm_area_struct **prev, unsigned long start, unsigned long end) madvise_remove() argument
|
H A D | slob.c | 219 slob_t *prev, *cur, *aligned = NULL; slob_page_alloc() local 222 for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) { slob_page_alloc() 236 prev = cur; slob_page_alloc() 243 if (prev) slob_page_alloc() 244 set_slob(prev, slob_units(prev), next); slob_page_alloc() 248 if (prev) slob_page_alloc() 249 set_slob(prev, slob_units(prev), cur + units); slob_page_alloc() 271 struct list_head *prev; slob_alloc() local 299 prev = sp->lru.prev; list_for_each_entry() 307 if (prev != slob_list->prev && list_for_each_entry() 308 slob_list->next != prev->next) list_for_each_entry() 309 list_move_tail(slob_list, prev->next); list_for_each_entry() 343 slob_t *prev, *next, *b = (slob_t *)block; slob_free() local 399 prev = sp->freelist; slob_free() 400 next = slob_next(prev); slob_free() 402 prev = next; slob_free() 403 next = slob_next(prev); slob_free() 406 if (!slob_last(prev) && b + units == next) { slob_free() 412 if (prev + slob_units(prev) == b) { slob_free() 413 units = slob_units(b) + slob_units(prev); slob_free() 414 set_slob(prev, units, slob_next(b)); slob_free() 416 set_slob(prev, slob_units(prev), b); slob_free()
|
H A D | mlock.c | 498 static int mlock_fixup(struct vm_area_struct *vma, struct vm_area_struct **prev, mlock_fixup() argument 512 *prev = vma_merge(mm, *prev, start, end, newflags, vma->anon_vma, mlock_fixup() 514 if (*prev) { mlock_fixup() 515 vma = *prev; mlock_fixup() 552 *prev = vma; mlock_fixup() 559 struct vm_area_struct * vma, * prev; do_mlock() local 573 prev = vma->vm_prev; do_mlock() 575 prev = vma; do_mlock() 589 error = mlock_fixup(vma, &prev, nstart, tmp, newflags); do_mlock() 593 if (nstart < prev->vm_end) do_mlock() 594 nstart = prev->vm_end; do_mlock() 598 vma = prev->vm_next; do_mlock() 659 struct vm_area_struct * vma, * prev = NULL; do_mlockall() local 668 for (vma = current->mm->mmap; vma ; vma = prev->vm_next) { do_mlockall() 676 mlock_fixup(vma, &prev, vma->vm_start, vma->vm_end, newflags); do_mlockall()
|
H A D | mmap.c | 61 struct vm_area_struct *vma, struct vm_area_struct *prev, 381 unsigned long prev = 0, pend = 0; browse_rb() local 386 if (vma->vm_start < prev) { browse_rb() 387 pr_emerg("vm_start %lx < prev %lx\n", browse_rb() 388 vma->vm_start, prev); browse_rb() 409 prev = vma->vm_start; browse_rb() 660 struct vm_area_struct *prev, struct rb_node **rb_link, __vma_link() 663 __vma_link_list(mm, vma, prev, rb_parent); __vma_link() 668 struct vm_area_struct *prev, struct rb_node **rb_link, vma_link() 678 __vma_link(mm, vma, prev, rb_link, rb_parent); vma_link() 694 struct vm_area_struct *prev; __insert_vm_struct() local 698 &prev, &rb_link, &rb_parent)) __insert_vm_struct() 700 __vma_link(mm, vma, prev, rb_link, rb_parent); __insert_vm_struct() 706 struct vm_area_struct *prev) __vma_unlink() 711 prev->vm_next = next = vma->vm_next; __vma_unlink() 713 next->vm_prev = prev; __vma_unlink() 1011 * an offset within prev, or at the start of next), and the flags of 1017 * vma, PPPPPP is the prev vma specified, and NNNNNN the next vma after: 1033 struct vm_area_struct *prev, unsigned long addr, vma_merge() 1049 if (prev) vma_merge() 1050 next = prev->vm_next; vma_merge() 1060 if (prev && prev->vm_end == addr && vma_merge() 1061 mpol_equal(vma_policy(prev), policy) && vma_merge() 1062 can_vma_merge_after(prev, vm_flags, vma_merge() 1071 is_mergeable_anon_vma(prev->anon_vma, vma_merge() 1074 err = vma_adjust(prev, prev->vm_start, vma_merge() 1075 next->vm_end, prev->vm_pgoff, NULL); vma_merge() 1077 err = vma_adjust(prev, prev->vm_start, vma_merge() 1078 end, prev->vm_pgoff, NULL); vma_merge() 1081 khugepaged_enter_vma_merge(prev, vm_flags); vma_merge() 1082 return prev; vma_merge() 1092 if (prev && addr < prev->vm_end) /* case 4 */ vma_merge() 1093 err = vma_adjust(prev, prev->vm_start, vma_merge() 1094 addr, prev->vm_pgoff, NULL); vma_merge() 1534 struct vm_area_struct *vma, *prev; mmap_region() local 1558 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, mmap_region() 1577 vma = vma_merge(mm, prev, addr, addr + len, vm_flags, NULL, file, pgoff, mmap_region() 1627 * Bug: If addr is changed, prev, rb_link, rb_parent should mmap_region() 1640 vma_link(mm, vma, prev, rb_link, rb_parent); mmap_region() 1682 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); mmap_region() 1766 struct rb_node *prev = &vma->vm_rb; unmapped_area() local 1767 if (!rb_parent(prev)) unmapped_area() 1769 vma = rb_entry(rb_parent(prev), unmapped_area() 1771 if (prev == vma->vm_rb.rb_left) { unmapped_area() 1869 struct rb_node *prev = &vma->vm_rb; unmapped_area_topdown() local 1870 if (!rb_parent(prev)) unmapped_area_topdown() 1872 vma = rb_entry(rb_parent(prev), unmapped_area_topdown() 1874 if (prev == vma->vm_rb.rb_right) { unmapped_area_topdown() 2305 struct vm_area_struct *vma, *prev; find_extend_vma() local 2308 vma = find_vma_prev(mm, addr, &prev); find_extend_vma() 2311 if (!prev || expand_stack(prev, addr)) find_extend_vma() 2313 if (prev->vm_flags & VM_LOCKED) find_extend_vma() 2314 populate_vma_page_range(prev, addr, prev->vm_end, NULL); find_extend_vma() 2315 return prev; find_extend_vma() 2320 struct vm_area_struct *prev; expand_stack() local 2323 prev = vma->vm_prev; expand_stack() 2324 if (prev && prev->vm_end == address) { expand_stack() 2325 if (!(prev->vm_flags & VM_GROWSDOWN)) expand_stack() 2386 struct vm_area_struct *vma, struct vm_area_struct *prev, unmap_region() 2389 struct vm_area_struct *next = prev ? prev->vm_next : mm->mmap; unmap_region() 2396 free_pgtables(&tlb, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, unmap_region() 2407 struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped() 2412 insertion_point = (prev ? &prev->vm_next : &mm->mmap); detach_vmas_to_be_unmapped() 2422 vma->vm_prev = prev; detach_vmas_to_be_unmapped() 2425 mm->highest_vm_end = prev ? prev->vm_end : 0; detach_vmas_to_be_unmapped() 2521 struct vm_area_struct *vma, *prev, *last; do_munmap() local 2534 prev = vma->vm_prev; do_munmap() 2563 prev = vma; do_munmap() 2573 vma = prev ? prev->vm_next : mm->mmap; do_munmap() 2592 detach_vmas_to_be_unmapped(mm, vma, prev, end); do_munmap() 2593 unmap_region(mm, vma, prev, start, end); do_munmap() 2732 struct vm_area_struct *vma, *prev; do_brk() local 2761 while (find_vma_links(mm, addr, addr + len, &prev, &rb_link, do_brk() 2778 vma = vma_merge(mm, prev, addr, addr + len, flags, do_brk() 2799 vma_link(mm, vma, prev, rb_link, rb_parent); do_brk() 2878 struct vm_area_struct *prev; insert_vm_struct() local 2898 &prev, &rb_link, &rb_parent)) insert_vm_struct() 2904 vma_link(mm, vma, prev, rb_link, rb_parent); insert_vm_struct() 2919 struct vm_area_struct *new_vma, *prev; copy_vma() local 2932 if (find_vma_links(mm, addr, addr + len, &prev, &rb_link, &rb_parent)) copy_vma() 2934 new_vma = vma_merge(mm, prev, addr, addr + len, vma->vm_flags, copy_vma() 2974 vma_link(mm, new_vma, prev, rb_link, rb_parent); copy_vma() 659 __vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) __vma_link() argument 667 vma_link(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, struct rb_node **rb_link, struct rb_node *rb_parent) vma_link() argument 705 __vma_unlink(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev) __vma_unlink() argument 1032 vma_merge(struct mm_struct *mm, struct vm_area_struct *prev, unsigned long addr, unsigned long end, unsigned long vm_flags, struct anon_vma *anon_vma, struct file *file, pgoff_t pgoff, struct mempolicy *policy) vma_merge() argument 2385 unmap_region(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long start, unsigned long end) unmap_region() argument 2406 detach_vmas_to_be_unmapped(struct mm_struct *mm, struct vm_area_struct *vma, struct vm_area_struct *prev, unsigned long end) detach_vmas_to_be_unmapped() argument
|
H A D | mprotect.c | 339 struct vm_area_struct *vma, *prev; SYSCALL_DEFINE3() local 372 prev = vma->vm_prev; SYSCALL_DEFINE3() 391 prev = vma; SYSCALL_DEFINE3() 414 error = mprotect_fixup(vma, &prev, nstart, tmp, newflags); SYSCALL_DEFINE3() 419 if (nstart < prev->vm_end) SYSCALL_DEFINE3() 420 nstart = prev->vm_end; SYSCALL_DEFINE3() 424 vma = prev->vm_next; SYSCALL_DEFINE3()
|
/linux-4.1.27/sound/pci/ctxfi/ |
H A D | ctimap.c | 46 pre = pos->prev; 48 pre = head->prev; 50 __list_add(&entry->list, pos->prev, pos); 52 pre = head->prev; 79 pre = (entry->list.prev == head) ? head->prev : entry->list.prev; input_mapper_delete()
|
/linux-4.1.27/arch/alpha/include/asm/ |
H A D | xchg.h | 132 * prev is equal to old) then we aren't acquiring anything new and 139 unsigned long prev, tmp, cmp, addr64; ____cmpxchg() local 157 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) ____cmpxchg() 160 return prev; ____cmpxchg() 166 unsigned long prev, tmp, cmp, addr64; ____cmpxchg() local 184 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) ____cmpxchg() 187 return prev; ____cmpxchg() 193 unsigned long prev, cmp; ____cmpxchg() local 207 : "=&r"(prev), "=&r"(cmp), "=m"(*m) ____cmpxchg() 210 return prev; ____cmpxchg() 216 unsigned long prev, cmp; ____cmpxchg() local 230 : "=&r"(prev), "=&r"(cmp), "=m"(*m) ____cmpxchg() 233 return prev; ____cmpxchg()
|
H A D | futex.h | 88 u32 prev; futex_atomic_cmpxchg_inatomic() local 110 : "+r"(ret), "=&r"(prev), "=&r"(cmp) futex_atomic_cmpxchg_inatomic() 114 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
H A D | gct.h | 24 gct6_handle prev; member in struct:__gct6_node
|
/linux-4.1.27/arch/s390/include/asm/ |
H A D | switch_to.h | 158 #define switch_to(prev,next,last) do { \ 159 if (prev->mm) { \ 160 save_fp_ctl(&prev->thread.fp_regs.fpc); \ 161 save_fp_vx_regs(prev); \ 162 save_access_regs(&prev->thread.acrs[0]); \ 163 save_ri_cb(prev->thread.ri_cb); \ 170 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \ 172 prev = __switch_to(prev,next); \
|
H A D | mmu_context.h | 69 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 75 if (prev == next) switch_mm() 83 atomic_dec(&prev->context.attach_count); switch_mm() 85 cpumask_clear_cpu(cpu, &prev->context.cpu_attach_mask); switch_mm() 111 static inline void activate_mm(struct mm_struct *prev, activate_mm() argument 114 switch_mm(prev, next, current); activate_mm()
|
/linux-4.1.27/arch/avr32/kernel/ |
H A D | switch_to.S | 15 /* Switch thread context from "prev" to "next", returning "last" 16 * r12 : prev 17 * r11 : &prev->thread + 1
|
/linux-4.1.27/tools/perf/util/include/linux/ |
H A D | list.h | 18 begin->prev->next = end->next; list_del_range() 19 end->next->prev = begin->prev; list_del_range()
|
/linux-4.1.27/security/selinux/ss/ |
H A D | hashtab.c | 42 struct hashtab_node *prev, *cur, *newnode; hashtab_insert() local 50 prev = NULL; hashtab_insert() 53 prev = cur; hashtab_insert() 65 if (prev) { hashtab_insert() 66 newnode->next = prev->next; hashtab_insert() 67 prev->next = newnode; hashtab_insert()
|
H A D | ebitmap.c | 51 struct ebitmap_node *n, *new, *prev; ebitmap_cpy() local 55 prev = NULL; ebitmap_cpy() 65 if (prev) ebitmap_cpy() 66 prev->next = new; ebitmap_cpy() 69 prev = new; ebitmap_cpy() 258 struct ebitmap_node *n, *prev, *new; ebitmap_set_bit() local 260 prev = NULL; ebitmap_set_bit() 281 if (prev) ebitmap_set_bit() 282 e->highbit = prev->startbit ebitmap_set_bit() 287 if (prev) ebitmap_set_bit() 288 prev->next = n->next; ebitmap_set_bit() 295 prev = n; ebitmap_set_bit() 313 if (prev) { ebitmap_set_bit() 314 new->next = prev->next; ebitmap_set_bit() 315 prev->next = new; ebitmap_set_bit()
|
H A D | mls.c | 37 int i, l, len, head, prev; mls_compute_context_len() local 52 prev = -2; mls_compute_context_len() 55 if (i - prev > 1) { ebitmap_for_each_positive_bit() 57 if (head != prev) { ebitmap_for_each_positive_bit() 58 nm = sym_name(&policydb, SYM_CATS, prev); ebitmap_for_each_positive_bit() 65 prev = i; ebitmap_for_each_positive_bit() 67 if (prev != head) { 68 nm = sym_name(&policydb, SYM_CATS, prev); 92 int i, l, head, prev; mls_sid_to_context() local 111 prev = -2; mls_sid_to_context() 114 if (i - prev > 1) { ebitmap_for_each_positive_bit() 116 if (prev != head) { ebitmap_for_each_positive_bit() 117 if (prev - head > 1) ebitmap_for_each_positive_bit() 121 nm = sym_name(&policydb, SYM_CATS, prev); ebitmap_for_each_positive_bit() 125 if (prev < 0) ebitmap_for_each_positive_bit() 134 prev = i; ebitmap_for_each_positive_bit() 137 if (prev != head) { 138 if (prev - head > 1) 142 nm = sym_name(&policydb, SYM_CATS, prev);
|
H A D | sidtab.c | 36 struct sidtab_node *prev, *cur, *newnode; sidtab_insert() local 44 prev = NULL; sidtab_insert() 47 prev = cur; sidtab_insert() 68 if (prev) { sidtab_insert() 69 newnode->next = prev->next; sidtab_insert() 71 prev->next = newnode; sidtab_insert()
|
H A D | avtab.c | 69 struct avtab_node *prev, struct avtab_node *cur, avtab_insert_node() 78 if (prev) { avtab_insert_node() 79 newnode->next = prev->next; avtab_insert_node() 80 prev->next = newnode; avtab_insert_node() 97 struct avtab_node *prev, *cur, *newnode; avtab_insert() local 104 for (prev = NULL, cur = flex_array_get_ptr(h->htable, hvalue); avtab_insert() 106 prev = cur, cur = cur->next) { avtab_insert() 123 newnode = avtab_insert_node(h, hvalue, prev, cur, key, datum); avtab_insert() 138 struct avtab_node *prev, *cur; avtab_insert_nonunique() local 144 for (prev = NULL, cur = flex_array_get_ptr(h->htable, hvalue); avtab_insert_nonunique() 146 prev = cur, cur = cur->next) { avtab_insert_nonunique() 162 return avtab_insert_node(h, hvalue, prev, cur, key, datum); avtab_insert_nonunique() 68 avtab_insert_node(struct avtab *h, int hvalue, struct avtab_node *prev, struct avtab_node *cur, struct avtab_key *key, struct avtab_datum *datum) avtab_insert_node() argument
|
/linux-4.1.27/fs/jffs2/ |
H A D | nodelist.c | 28 struct jffs2_full_dirent **prev = list; jffs2_add_fd_to_list() local 32 while ((*prev) && (*prev)->nhash <= new->nhash) { jffs2_add_fd_to_list() 33 if ((*prev)->nhash == new->nhash && !strcmp((*prev)->name, new->name)) { jffs2_add_fd_to_list() 35 if (new->version < (*prev)->version) { jffs2_add_fd_to_list() 37 (*prev)->name, (*prev)->ino); jffs2_add_fd_to_list() 42 (*prev)->name, (*prev)->ino); jffs2_add_fd_to_list() 43 new->next = (*prev)->next; jffs2_add_fd_to_list() 46 if ((*prev)->raw) jffs2_add_fd_to_list() 47 jffs2_mark_node_obsolete(c, ((*prev)->raw)); jffs2_add_fd_to_list() 48 jffs2_free_full_dirent(*prev); jffs2_add_fd_to_list() 49 *prev = new; jffs2_add_fd_to_list() 53 prev = &((*prev)->next); jffs2_add_fd_to_list() 55 new->next = *prev; jffs2_add_fd_to_list() 56 *prev = new; jffs2_add_fd_to_list() 386 struct jffs2_node_frag *prev = frag_prev(newfrag); jffs2_add_full_dnode_to_inode() local 390 if (prev->node) jffs2_add_full_dnode_to_inode() 391 mark_ref_normal(prev->node->raw); jffs2_add_full_dnode_to_inode() 438 struct jffs2_inode_cache **prev; jffs2_add_ino_cache() local 446 prev = &c->inocache_list[new->ino % c->inocache_hashsize]; jffs2_add_ino_cache() 448 while ((*prev) && (*prev)->ino < new->ino) { jffs2_add_ino_cache() 449 prev = &(*prev)->next; jffs2_add_ino_cache() 451 new->next = *prev; jffs2_add_ino_cache() 452 *prev = new; jffs2_add_ino_cache() 459 struct jffs2_inode_cache **prev; jffs2_del_ino_cache() local 467 prev = &c->inocache_list[old->ino % c->inocache_hashsize]; jffs2_del_ino_cache() 469 while ((*prev) && (*prev)->ino < old->ino) { jffs2_del_ino_cache() 470 prev = &(*prev)->next; jffs2_del_ino_cache() 472 if ((*prev) == old) { jffs2_del_ino_cache() 473 *prev = old->next; jffs2_del_ino_cache() 529 struct jffs2_node_frag *prev = NULL; jffs2_lookup_node_frag() local 541 if (!prev || frag->ofs > prev->ofs) jffs2_lookup_node_frag() 542 prev = frag; jffs2_lookup_node_frag() 554 if (prev) jffs2_lookup_node_frag() 556 prev->ofs, prev->ofs+prev->size); jffs2_lookup_node_frag() 560 return prev; jffs2_lookup_node_frag()
|
H A D | erase.c | 239 struct jffs2_raw_node_ref **prev; jffs2_remove_node_refs_from_ino_list() local 241 prev = &ref->next_in_ino; jffs2_remove_node_refs_from_ino_list() 245 if (!(*prev)->next_in_ino) { jffs2_remove_node_refs_from_ino_list() 249 ic = (struct jffs2_inode_cache *)(*prev); jffs2_remove_node_refs_from_ino_list() 250 prev = &ic->nodes; jffs2_remove_node_refs_from_ino_list() 254 if (SECTOR_ADDR((*prev)->flash_offset) == jeb->offset) { jffs2_remove_node_refs_from_ino_list() 258 this = *prev; jffs2_remove_node_refs_from_ino_list() 259 *prev = this->next_in_ino; jffs2_remove_node_refs_from_ino_list() 268 prev = &((*prev)->next_in_ino); jffs2_remove_node_refs_from_ino_list()
|
/linux-4.1.27/drivers/video/fbdev/omap2/dss/ |
H A D | dss-of.c | 27 struct device_node *prev) omapdss_of_get_next_port() 34 if (!prev) { omapdss_of_get_next_port() 51 ports = of_get_parent(prev); omapdss_of_get_next_port() 56 port = of_get_next_child(ports, prev); omapdss_of_get_next_port() 61 prev = port; omapdss_of_get_next_port() 71 struct device_node *prev) omapdss_of_get_next_endpoint() 79 ep = of_get_next_child(parent, prev); omapdss_of_get_next_endpoint() 82 prev = ep; omapdss_of_get_next_endpoint() 26 omapdss_of_get_next_port(const struct device_node *parent, struct device_node *prev) omapdss_of_get_next_port() argument 70 omapdss_of_get_next_endpoint(const struct device_node *parent, struct device_node *prev) omapdss_of_get_next_endpoint() argument
|
/linux-4.1.27/arch/nios2/include/asm/ |
H A D | switch_to.h | 17 #define switch_to(prev, next, last) \ 26 : "r" (prev), "r" (next) \
|
H A D | mmu_context.h | 52 void switch_mm(struct mm_struct *prev, struct mm_struct *next, 64 void activate_mm(struct mm_struct *prev, struct mm_struct *next);
|
/linux-4.1.27/fs/btrfs/ |
H A D | extent_map.c | 145 struct rb_node *prev = NULL; __tree_search() local 152 prev = n; __tree_search() 164 orig_prev = prev; __tree_search() 165 while (prev && offset >= extent_map_end(prev_entry)) { __tree_search() 166 prev = rb_next(prev); __tree_search() 167 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search() 169 *prev_ret = prev; __tree_search() 170 prev = orig_prev; __tree_search() 174 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search() 175 while (prev && offset < prev_entry->start) { __tree_search() 176 prev = rb_prev(prev); __tree_search() 177 prev_entry = rb_entry(prev, struct extent_map, rb_node); __tree_search() 179 *next_ret = prev; __tree_search() 185 static int mergable_maps(struct extent_map *prev, struct extent_map *next) mergable_maps() argument 187 if (test_bit(EXTENT_FLAG_PINNED, &prev->flags)) mergable_maps() 194 if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags)) mergable_maps() 197 if (test_bit(EXTENT_FLAG_LOGGING, &prev->flags) || mergable_maps() 206 if (!list_empty(&prev->list) || !list_empty(&next->list)) mergable_maps() 209 if (extent_map_end(prev) == next->start && mergable_maps() 210 prev->flags == next->flags && mergable_maps() 211 prev->bdev == next->bdev && mergable_maps() 213 prev->block_start == EXTENT_MAP_HOLE) || mergable_maps() 215 prev->block_start == EXTENT_MAP_INLINE) || mergable_maps() 217 prev->block_start == EXTENT_MAP_DELALLOC) || mergable_maps() 219 next->block_start == extent_map_block_end(prev)))) { mergable_maps() 365 struct rb_node *prev = NULL; __lookup_extent_mapping() local 369 rb_node = __tree_search(&tree->map, start, &prev, &next); __lookup_extent_mapping() 371 if (prev) __lookup_extent_mapping() 372 rb_node = prev; __lookup_extent_mapping()
|
H A D | ordered-data.c | 81 struct rb_node *prev = NULL; __tree_search() local 88 prev = n; __tree_search() 101 while (prev && file_offset >= entry_end(prev_entry)) { __tree_search() 102 test = rb_next(prev); __tree_search() 110 prev = test; __tree_search() 112 if (prev) __tree_search() 113 prev_entry = rb_entry(prev, struct btrfs_ordered_extent, __tree_search() 115 while (prev && file_offset < entry_end(prev_entry)) { __tree_search() 116 test = rb_prev(prev); __tree_search() 121 prev = test; __tree_search() 123 *prev_ret = prev; __tree_search() 155 struct rb_node *prev = NULL; tree_search() local 165 ret = __tree_search(root, file_offset, &prev); tree_search() 167 ret = prev; tree_search() 442 struct rb_node *prev; btrfs_get_logged_extents() local 446 n = __tree_search(&tree->tree, end, &prev); btrfs_get_logged_extents() 448 n = prev; btrfs_get_logged_extents() 883 struct rb_node *prev = NULL; btrfs_ordered_update_i_size() local 929 prev = tree_search(tree, offset); btrfs_ordered_update_i_size() 934 if (prev) { btrfs_ordered_update_i_size() 935 test = rb_entry(prev, struct btrfs_ordered_extent, btrfs_ordered_update_i_size() 939 node = prev; btrfs_ordered_update_i_size()
|
/linux-4.1.27/include/uapi/linux/netfilter_bridge/ |
H A D | ebt_limit.h | 19 unsigned long prev; member in struct:ebt_limit_info
|
/linux-4.1.27/drivers/usb/host/ |
H A D | ohci-mem.c | 105 struct td **prev = &hc->td_hash [TD_HASH_FUNC (td->td_dma)]; td_free() local 107 while (*prev && *prev != td) td_free() 108 prev = &(*prev)->td_hash; td_free() 109 if (*prev) td_free() 110 *prev = td->td_hash; td_free()
|
/linux-4.1.27/arch/m68k/include/asm/ |
H A D | atomic.h | 128 int prev; atomic_cmpxchg() local 131 prev = atomic_read(v); atomic_cmpxchg() 132 if (prev == old) atomic_cmpxchg() 135 return prev; atomic_cmpxchg() 141 int prev; atomic_xchg() local 144 prev = atomic_read(v); atomic_xchg() 147 return prev; atomic_xchg()
|
H A D | switch_to.h | 27 * pass prev in a0, next in a1 30 #define switch_to(prev,next,last) do { \ 31 register void *_prev __asm__ ("a0") = (prev); \
|
/linux-4.1.27/drivers/gpu/drm/nouveau/include/nvif/ |
H A D | list.h | 111 struct list_head *next, *prev; member in struct:list_head 130 list->next = list->prev = list; INIT_LIST_HEAD() 135 struct list_head *prev, struct list_head *next) __list_add() 137 next->prev = entry; __list_add() 139 entry->prev = prev; __list_add() 140 prev->next = entry; __list_add() 182 __list_add(entry, head->prev, head); list_add_tail() 186 __list_del(struct list_head *prev, struct list_head *next) __list_del() argument 188 next->prev = prev; __list_del() 189 prev->next = next; __list_del() 209 __list_del(entry->prev, entry->next); list_del() 215 __list_del(entry->prev, entry->next); list_del_init() 222 __list_del(list->prev, list->next); list_move_tail() 292 list_entry((ptr)->prev, type, member) 334 for (pos = __container_of((head)->prev, pos, member); \ 336 pos = __container_of(pos->member.prev, pos, member)) 344 for (pos = __container_of(pos->member.prev, pos, member); \ 346 pos = __container_of(pos->member.prev, pos, member)) 134 __list_add(struct list_head *entry, struct list_head *prev, struct list_head *next) __list_add() argument
|
/linux-4.1.27/arch/hexagon/include/asm/ |
H A D | mmu_context.h | 66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 75 if (next->context.generation < prev->context.generation) { switch_mm() 79 next->context.generation = prev->context.generation; switch_mm() 88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) activate_mm() argument 93 switch_mm(prev, next, current_thread_info()->task); activate_mm()
|
H A D | futex.h | 106 int prev; futex_atomic_cmpxchg_inatomic() local 128 : "+r" (ret), "=&r" (prev), "+m" (*uaddr) futex_atomic_cmpxchg_inatomic() 132 *uval = prev; futex_atomic_cmpxchg_inatomic()
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
H A D | mthca_mcg.c | 54 * If GID is found in MGM or MGM is empty, *index = *hash, *prev = -1 57 * if GID is found in AMGM, *index = index in AMGM, *prev = index of 60 * If no AMGM exists for given gid, *index = -1, *prev = index of last 65 u16 *hash, int *prev, int *index) find_mgm() 89 *prev = -1; find_mgm() 109 *prev = *index; find_mgm() 126 int index, prev; mthca_multicast_attach() local 138 err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); mthca_multicast_attach() 191 err = mthca_READ_MGM(dev, prev, mailbox); mthca_multicast_attach() 199 err = mthca_WRITE_MGM(dev, prev, mailbox); mthca_multicast_attach() 220 int prev, index; mthca_multicast_detach() local 231 err = find_mgm(dev, gid->raw, mailbox, &hash, &prev, &index); mthca_multicast_detach() 266 if (prev == -1) { mthca_multicast_detach() 291 err = mthca_READ_MGM(dev, prev, mailbox); mthca_multicast_detach() 299 err = mthca_WRITE_MGM(dev, prev, mailbox); mthca_multicast_detach() 63 find_mgm(struct mthca_dev *dev, u8 *gid, struct mthca_mailbox *mgm_mailbox, u16 *hash, int *prev, int *index) find_mgm() argument
|
/linux-4.1.27/tools/perf/scripts/python/ |
H A D | sched-migration.py | 102 def sched_switch(self, prev, prev_state, next): 106 and prev in self.tasks: 110 event = RunqueueEventSleep(prev) 113 if prev in self.tasks: 115 next_tasks.remove(prev) 117 next_tasks.append(prev) 160 def __init__(self, start, prev): 162 self.prev = prev 166 if prev is not None: 167 self.total_load = prev.total_load 168 self.rqs = prev.rqs.copy() 177 def sched_switch(self, ts_list, prev, prev_state, next, cpu): 178 old_rq = self.prev.rqs[cpu] 179 new_rq = old_rq.sched_switch(prev, prev_state, next) 192 old_rq = self.prev.rqs[old_cpu] 197 new_rq = self.prev.rqs[new_cpu] 209 old_rq = self.prev.rqs[cpu] 339 print "Sched switch event rejected ts: %s cpu: %d prev: %s(%d) next: %s(%d)" % \
|
/linux-4.1.27/drivers/gpu/drm/nouveau/ |
H A D | nv17_fence.c | 34 struct nouveau_channel *prev, struct nouveau_channel *chan) nv17_fence_sync() 36 struct nouveau_cli *cli = (void *)nvif_client(&prev->device->base); nv17_fence_sync() 50 ret = RING_SPACE(prev, 5); nv17_fence_sync() 52 BEGIN_NV04(prev, 0, NV11_SUBCHAN_DMA_SEMAPHORE, 4); nv17_fence_sync() 53 OUT_RING (prev, fctx->sema.handle); nv17_fence_sync() 54 OUT_RING (prev, 0); nv17_fence_sync() 55 OUT_RING (prev, value + 0); nv17_fence_sync() 56 OUT_RING (prev, value + 1); nv17_fence_sync() 57 FIRE_RING (prev); nv17_fence_sync() 33 nv17_fence_sync(struct nouveau_fence *fence, struct nouveau_channel *prev, struct nouveau_channel *chan) nv17_fence_sync() argument
|
H A D | nouveau_bo.h | 53 struct nouveau_bo *prev; nouveau_bo_ref() local 57 prev = *pnvbo; nouveau_bo_ref() 60 if (prev) { nouveau_bo_ref() 61 struct ttm_buffer_object *bo = &prev->bo; nouveau_bo_ref()
|
H A D | nouveau_fence.c | 411 struct nouveau_channel *prev = NULL; nouveau_fence_sync() local 417 prev = rcu_dereference(f->channel); nouveau_fence_sync() 418 if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) nouveau_fence_sync() 433 struct nouveau_channel *prev = NULL; nouveau_fence_sync() local 442 prev = rcu_dereference(f->channel); nouveau_fence_sync() 443 if (prev && (prev == chan || fctx->sync(f, prev, chan) == 0)) nouveau_fence_sync()
|
/linux-4.1.27/drivers/char/agp/ |
H A D | frontend.c | 68 struct agp_memory *prev; agp_remove_from_pool() local 76 prev = temp->prev; agp_remove_from_pool() 78 if (prev != NULL) { agp_remove_from_pool() 79 prev->next = next; agp_remove_from_pool() 81 next->prev = prev; agp_remove_from_pool() 86 next->prev = NULL; agp_remove_from_pool() 201 struct agp_memory *prev; agp_insert_into_pool() local 203 prev = agp_fe.current_controller->pool; agp_insert_into_pool() 205 if (prev != NULL) { agp_insert_into_pool() 206 prev->prev = temp; agp_insert_into_pool() 207 temp->next = prev; agp_insert_into_pool() 232 struct agp_file_private *prev; agp_insert_file_private() local 234 prev = agp_fe.file_priv_list; agp_insert_file_private() 236 if (prev != NULL) agp_insert_file_private() 237 prev->prev = priv; agp_insert_file_private() 238 priv->next = prev; agp_insert_file_private() 245 struct agp_file_private *prev; agp_remove_file_private() local 248 prev = priv->prev; agp_remove_file_private() 250 if (prev != NULL) { agp_remove_file_private() 251 prev->next = next; agp_remove_file_private() 254 next->prev = prev; agp_remove_file_private() 258 next->prev = NULL; agp_remove_file_private() 328 prev_controller->prev = controller; agp_insert_controller() 377 prev_controller = controller->prev; agp_remove_controller() 383 next_controller->prev = prev_controller; agp_remove_controller() 387 next_controller->prev = NULL; agp_remove_controller() 507 prev_client->prev = client; agp_insert_client() 541 prev_client = client->prev; agp_remove_client() 547 next_client->prev = prev_client; agp_remove_client() 551 next_client->prev = NULL; agp_remove_client()
|
/linux-4.1.27/crypto/ |
H A D | cmac.c | 46 * | prev (block size) 113 u8 *prev = PTR_ALIGN((void *)ctx->ctx, alignmask + 1) + bs; crypto_cmac_digest_init() local 116 memset(prev, 0, bs); crypto_cmac_digest_init() 131 u8 *prev = odds + bs; crypto_cmac_digest_update() local 145 crypto_xor(prev, odds, bs); crypto_cmac_digest_update() 146 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_cmac_digest_update() 153 crypto_xor(prev, p, bs); crypto_cmac_digest_update() 154 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_cmac_digest_update() 178 u8 *prev = odds + bs; crypto_cmac_digest_final() local 195 crypto_xor(prev, odds, bs); crypto_cmac_digest_final() 196 crypto_xor(prev, consts + offset, bs); crypto_cmac_digest_final() 198 crypto_cipher_encrypt_one(tfm, out, prev); crypto_cmac_digest_final()
|
H A D | xcbc.c | 53 * | prev (block size) 87 u8 *prev = PTR_ALIGN(&ctx->ctx[0], alignmask + 1) + bs; crypto_xcbc_digest_init() local 90 memset(prev, 0, bs); crypto_xcbc_digest_init() 105 u8 *prev = odds + bs; crypto_xcbc_digest_update() local 119 crypto_xor(prev, odds, bs); crypto_xcbc_digest_update() 120 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_xcbc_digest_update() 127 crypto_xor(prev, p, bs); crypto_xcbc_digest_update() 128 crypto_cipher_encrypt_one(tfm, prev, prev); crypto_xcbc_digest_update() 152 u8 *prev = odds + bs; crypto_xcbc_digest_final() local 169 crypto_xor(prev, odds, bs); crypto_xcbc_digest_final() 170 crypto_xor(prev, consts + offset, bs); crypto_xcbc_digest_final() 172 crypto_cipher_encrypt_one(tfm, out, prev); crypto_xcbc_digest_final()
|
/linux-4.1.27/sound/isa/gus/ |
H A D | gus_mem.c | 54 nblock->prev = pblock->prev; snd_gf1_mem_xalloc() 56 pblock->prev = nblock; snd_gf1_mem_xalloc() 60 nblock->prev->next = nblock; snd_gf1_mem_xalloc() 68 nblock->prev = NULL; snd_gf1_mem_xalloc() 71 nblock->prev = alloc->last; snd_gf1_mem_xalloc() 88 block->next->prev = NULL; snd_gf1_mem_xfree() 90 block->prev->next = block->next; snd_gf1_mem_xfree() 92 block->next->prev = block->prev; snd_gf1_mem_xfree() 95 alloc->last = block->prev; snd_gf1_mem_xfree() 96 if (block->prev) snd_gf1_mem_xfree() 97 block->prev->next = NULL; snd_gf1_mem_xfree() 99 block->next->prev = block->prev; snd_gf1_mem_xfree() 100 if (block->prev) snd_gf1_mem_xfree() 101 block->prev->next = block->next; snd_gf1_mem_xfree() 154 block->prev = block->next = NULL; snd_gf1_mem_find()
|
/linux-4.1.27/arch/sparc/lib/ |
H A D | atomic32.c | 143 u32 prev; __cmpxchg_u32() local 146 if ((prev = *ptr) == old) __cmpxchg_u32() 150 return (unsigned long)prev; __cmpxchg_u32() 157 u32 prev; __xchg_u32() local 160 prev = *ptr; __xchg_u32() 164 return (unsigned long)prev; __xchg_u32()
|
H A D | memcpy.S | 100 #define SMOVE_CHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ 106 or %t5, %prev, %t5; \ 107 sll %t1, shil, %prev; \ 112 or %t1, %prev, %t1; \ 115 sll %t3, shil, %prev; \ 118 #define SMOVE_ALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, prev, shil, shir, offset2) \ 124 or %t4, %prev, %t0; \ 125 sll %t1, shil, %prev; \ 130 or %t4, %prev, %t2; \ 131 sll %t3, shil, %prev; \
|
/linux-4.1.27/arch/powerpc/platforms/cell/spufs/ |
H A D | switch.c | 1791 static int quiece_spu(struct spu_state *prev, struct spu *spu) quiece_spu() argument 1803 if (check_spu_isolate(prev, spu)) { /* Step 2. */ quiece_spu() 1806 disable_interrupts(prev, spu); /* Step 3. */ quiece_spu() 1807 set_watchdog_timer(prev, spu); /* Step 4. */ quiece_spu() 1808 inhibit_user_access(prev, spu); /* Step 5. */ quiece_spu() 1809 if (check_spu_isolate(prev, spu)) { /* Step 6. */ quiece_spu() 1812 set_switch_pending(prev, spu); /* Step 7. */ quiece_spu() 1813 save_mfc_cntl(prev, spu); /* Step 8. */ quiece_spu() 1814 save_spu_runcntl(prev, spu); /* Step 9. */ quiece_spu() 1815 save_mfc_sr1(prev, spu); /* Step 10. */ quiece_spu() 1816 save_spu_status(prev, spu); /* Step 11. */ quiece_spu() 1817 save_mfc_stopped_status(prev, spu); /* Step 12. */ quiece_spu() 1818 halt_mfc_decr(prev, spu); /* Step 13. */ quiece_spu() 1819 save_timebase(prev, spu); /* Step 14. */ quiece_spu() 1820 remove_other_spu_access(prev, spu); /* Step 15. */ quiece_spu() 1821 do_mfc_mssync(prev, spu); /* Step 16. */ quiece_spu() 1822 issue_mfc_tlbie(prev, spu); /* Step 17. */ quiece_spu() 1823 handle_pending_interrupts(prev, spu); /* Step 18. */ quiece_spu() 1828 static void save_csa(struct spu_state *prev, struct spu *spu) save_csa() argument 1835 save_mfc_queues(prev, spu); /* Step 19. */ save_csa() 1836 save_ppu_querymask(prev, spu); /* Step 20. */ save_csa() 1837 save_ppu_querytype(prev, spu); /* Step 21. */ save_csa() 1838 save_ppu_tagstatus(prev, spu); /* NEW. */ save_csa() 1839 save_mfc_csr_tsq(prev, spu); /* Step 22. */ save_csa() 1840 save_mfc_csr_cmd(prev, spu); /* Step 23. */ save_csa() 1841 save_mfc_csr_ato(prev, spu); /* Step 24. */ save_csa() 1842 save_mfc_tclass_id(prev, spu); /* Step 25. */ save_csa() 1843 set_mfc_tclass_id(prev, spu); /* Step 26. */ save_csa() 1844 save_mfc_cmd(prev, spu); /* Step 26a - moved from 44. */ save_csa() 1845 purge_mfc_queue(prev, spu); /* Step 27. */ save_csa() 1846 wait_purge_complete(prev, spu); /* Step 28. */ save_csa() 1847 setup_mfc_sr1(prev, spu); /* Step 30. */ save_csa() 1848 save_spu_npc(prev, spu); /* Step 31. */ save_csa() 1849 save_spu_privcntl(prev, spu); /* Step 32. */ save_csa() 1850 reset_spu_privcntl(prev, spu); /* Step 33. */ save_csa() 1851 save_spu_lslr(prev, spu); /* Step 34. */ save_csa() 1852 reset_spu_lslr(prev, spu); /* Step 35. */ save_csa() 1853 save_spu_cfg(prev, spu); /* Step 36. */ save_csa() 1854 save_pm_trace(prev, spu); /* Step 37. */ save_csa() 1855 save_mfc_rag(prev, spu); /* Step 38. */ save_csa() 1856 save_ppu_mb_stat(prev, spu); /* Step 39. */ save_csa() 1857 save_ppu_mb(prev, spu); /* Step 40. */ save_csa() 1858 save_ppuint_mb(prev, spu); /* Step 41. */ save_csa() 1859 save_ch_part1(prev, spu); /* Step 42. */ save_csa() 1860 save_spu_mb(prev, spu); /* Step 43. */ save_csa() 1861 reset_ch(prev, spu); /* Step 45. */ save_csa() 1864 static void save_lscsa(struct spu_state *prev, struct spu *spu) save_lscsa() argument 1872 resume_mfc_queue(prev, spu); /* Step 46. */ save_lscsa() 1874 setup_mfc_slbs(prev, spu, spu_save_code, sizeof(spu_save_code)); save_lscsa() 1875 set_switch_active(prev, spu); /* Step 48. */ save_lscsa() 1876 enable_interrupts(prev, spu); /* Step 49. */ save_lscsa() 1877 save_ls_16kb(prev, spu); /* Step 50. */ save_lscsa() 1878 set_spu_npc(prev, spu); /* Step 51. */ save_lscsa() 1879 set_signot1(prev, spu); /* Step 52. */ save_lscsa() 1880 set_signot2(prev, spu); /* Step 53. */ save_lscsa() 1881 send_save_code(prev, spu); /* Step 54. */ save_lscsa() 1882 set_ppu_querymask(prev, spu); /* Step 55. */ save_lscsa() 1883 wait_tag_complete(prev, spu); /* Step 56. */ save_lscsa() 1884 wait_spu_stopped(prev, spu); /* Step 57. */ save_lscsa() 1932 static void harvest(struct spu_state *prev, struct spu *spu) harvest() argument 1940 disable_interrupts(prev, spu); /* Step 2. */ harvest() 1941 inhibit_user_access(prev, spu); /* Step 3. */ harvest() 1942 terminate_spu_app(prev, spu); /* Step 4. */ harvest() 1943 set_switch_pending(prev, spu); /* Step 5. */ harvest() 1945 remove_other_spu_access(prev, spu); /* Step 6. */ harvest() 1946 suspend_mfc_and_halt_decr(prev, spu); /* Step 7. */ harvest() 1947 wait_suspend_mfc_complete(prev, spu); /* Step 8. */ harvest() 1948 if (!suspend_spe(prev, spu)) /* Step 9. */ harvest() 1949 clear_spu_status(prev, spu); /* Step 10. */ harvest() 1950 do_mfc_mssync(prev, spu); /* Step 11. */ harvest() 1951 issue_mfc_tlbie(prev, spu); /* Step 12. */ harvest() 1952 handle_pending_interrupts(prev, spu); /* Step 13. */ harvest() 1953 purge_mfc_queue(prev, spu); /* Step 14. */ harvest() 1954 wait_purge_complete(prev, spu); /* Step 15. */ harvest() 1955 reset_spu_privcntl(prev, spu); /* Step 16. */ harvest() 1956 reset_spu_lslr(prev, spu); /* Step 17. */ harvest() 1957 setup_mfc_sr1(prev, spu); /* Step 18. */ harvest() 1959 reset_ch_part1(prev, spu); /* Step 20. */ harvest() 1960 reset_ch_part2(prev, spu); /* Step 21. */ harvest() 1961 enable_interrupts(prev, spu); /* Step 22. */ harvest() 1962 set_switch_active(prev, spu); /* Step 23. */ harvest() 1963 set_mfc_tclass_id(prev, spu); /* Step 24. */ harvest() 1964 resume_mfc_queue(prev, spu); /* Step 25. */ harvest() 2038 static int __do_spu_save(struct spu_state *prev, struct spu *spu) __do_spu_save() argument 2054 rc = quiece_spu(prev, spu); /* Steps 2-16. */ __do_spu_save() 2059 harvest(prev, spu); __do_spu_save() 2065 save_csa(prev, spu); /* Steps 17-43. */ __do_spu_save() 2066 save_lscsa(prev, spu); /* Steps 44-53. */ __do_spu_save() 2067 return check_save_status(prev, spu); /* Step 54. */ __do_spu_save() 2103 * @prev: pointer to SPU context save area, to be saved. 2108 int spu_save(struct spu_state *prev, struct spu *spu) spu_save() argument 2113 rc = __do_spu_save(prev, spu); /* Steps 2-53. */ spu_save()
|
/linux-4.1.27/arch/s390/lib/ |
H A D | spinlock.c | 186 void _raw_write_lock_wait(arch_rwlock_t *rw, unsigned int prev) _raw_write_lock_wait() argument 202 prev = __RAW_LOCK(&rw->lock, 0x80000000, __RAW_OP_OR); _raw_write_lock_wait() 203 old = prev; _raw_write_lock_wait() 205 if ((old & 0x7fffffff) == 0 && (int) prev >= 0) _raw_write_lock_wait() 217 unsigned int owner, old, prev; _raw_write_lock_wait() local 220 prev = 0x80000000; _raw_write_lock_wait() 232 prev = old; _raw_write_lock_wait() 235 if ((old & 0x7fffffff) == 0 && (int) prev >= 0) _raw_write_lock_wait()
|
/linux-4.1.27/tools/perf/ |
H A D | perf-completion.sh | 73 prev) 74 prev=${words_[$cword_-1]} 162 elif [[ $prev == @("-e"|"--event") && 232 local _ret=1 cur cword prev 234 prev=${words[CURRENT-1]} 248 local cur words cword prev 250 _get_comp_words_by_ref -n =: cur words cword prev 252 __perf_get_comp_words_by_ref -n =: cur words cword prev
|
H A D | builtin-timechart.c | 725 struct io_sample *prev; pid_begin_io_sample() local 736 prev = c->io_samples; pid_begin_io_sample() 738 if (prev && prev->start_time && !prev->end_time) { pid_begin_io_sample() 744 c->io_samples = prev->next; pid_begin_io_sample() 745 free(prev); pid_begin_io_sample() 769 struct io_sample *sample, *prev; pid_end_io_sample() local 793 prev = sample->next; pid_end_io_sample() 799 if (prev && sample->start_time < prev->end_time) { pid_end_io_sample() 800 if (prev->err) /* try to make errors more visible */ pid_end_io_sample() 801 sample->start_time = prev->end_time; pid_end_io_sample() 803 prev->end_time = sample->start_time; pid_end_io_sample() 820 if (prev && pid_end_io_sample() 821 prev->type == sample->type && pid_end_io_sample() 822 prev->err == sample->err && pid_end_io_sample() 823 prev->fd == sample->fd && pid_end_io_sample() 824 prev->end_time + tchart->merge_dist >= sample->start_time) { pid_end_io_sample() 826 sample->bytes += prev->bytes; pid_end_io_sample() 827 sample->merges += prev->merges + 1; pid_end_io_sample() 829 sample->start_time = prev->start_time; pid_end_io_sample() 830 sample->next = prev->next; pid_end_io_sample() 831 free(prev); pid_end_io_sample() 967 struct per_pid *new_list, *p, *cursor, *prev; sort_pids() local 982 prev = NULL; sort_pids() 988 if (prev) { sort_pids() 989 p->next = prev->next; sort_pids() 990 prev->next = p; sort_pids() 1001 prev = cursor; sort_pids() 1004 prev->next = p; sort_pids()
|
/linux-4.1.27/include/uapi/linux/netfilter/ |
H A D | xt_limit.h | 18 unsigned long prev; /* moved to xt_limit_priv */ member in struct:xt_rateinfo
|
/linux-4.1.27/arch/um/include/asm/ |
H A D | mmu_context.h | 50 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 55 if(prev != next){ switch_mm() 56 cpumask_clear_cpu(cpu, mm_cpumask(prev)); switch_mm()
|
/linux-4.1.27/arch/openrisc/include/asm/ |
H A D | mmu_context.h | 26 extern void switch_mm(struct mm_struct *prev, struct mm_struct *next, 31 #define activate_mm(prev, next) switch_mm((prev), (next), NULL)
|
/linux-4.1.27/kernel/sched/ |
H A D | idle_task.c | 27 pick_next_task_idle(struct rq *rq, struct task_struct *prev) pick_next_task_idle() argument 29 put_prev_task(rq, prev); pick_next_task_idle() 48 static void put_prev_task_idle(struct rq *rq, struct task_struct *prev) put_prev_task_idle() argument
|
H A D | stop_task.c | 27 pick_next_task_stop(struct rq *rq, struct task_struct *prev) pick_next_task_stop() argument 34 put_prev_task(rq, prev); pick_next_task_stop() 58 static void put_prev_task_stop(struct rq *rq, struct task_struct *prev) put_prev_task_stop() argument
|
H A D | stats.h | 128 * the idle task.) We are only called when prev != next. 132 struct task_struct *prev, struct task_struct *next) __sched_info_switch() 135 * prev now departs the cpu. It's not interesting to record __sched_info_switch() 139 if (prev != rq->idle) __sched_info_switch() 140 sched_info_depart(rq, prev); __sched_info_switch() 147 struct task_struct *prev, struct task_struct *next) sched_info_switch() 150 __sched_info_switch(rq, prev, next); sched_info_switch() 131 __sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) __sched_info_switch() argument 146 sched_info_switch(struct rq *rq, struct task_struct *prev, struct task_struct *next) sched_info_switch() argument
|
H A D | cputime.c | 392 void vtime_common_task_switch(struct task_struct *prev) vtime_common_task_switch() argument 394 if (is_idle_task(prev)) vtime_common_task_switch() 395 vtime_account_idle(prev); vtime_common_task_switch() 397 vtime_account_system(prev); vtime_common_task_switch() 400 vtime_account_user(prev); vtime_common_task_switch() 402 arch_vtime_task_switch(prev); vtime_common_task_switch() 579 struct cputime *prev, cputime_adjust() 601 if (prev->stime + prev->utime >= rtime) cputime_adjust() 619 cputime_advance(&prev->stime, stime); cputime_adjust() 620 cputime_advance(&prev->utime, utime); cputime_adjust() 623 *ut = prev->utime; cputime_adjust() 624 *st = prev->stime; cputime_adjust() 743 void arch_vtime_task_switch(struct task_struct *prev) arch_vtime_task_switch() argument 745 write_seqlock(&prev->vtime_seqlock); arch_vtime_task_switch() 746 prev->vtime_snap_whence = VTIME_SLEEPING; arch_vtime_task_switch() 747 write_sequnlock(&prev->vtime_seqlock); arch_vtime_task_switch() 578 cputime_adjust(struct task_cputime *curr, struct cputime *prev, cputime_t *ut, cputime_t *st) cputime_adjust() argument
|
H A D | auto_group.c | 130 struct autogroup *prev; autogroup_move_group() local 136 prev = p->signal->autogroup; autogroup_move_group() 137 if (prev == ag) { autogroup_move_group() 151 autogroup_kref_put(prev); autogroup_move_group()
|
/linux-4.1.27/drivers/isdn/divert/ |
H A D | isdn_divert.c | 30 struct call_struc *prev; member in struct:call_struc 38 struct deflect_struc *next, *prev; member in struct:deflect_struc 93 if (cs->prev) deflect_timer_expire() 94 cs->prev->next = cs->next; /* forward link */ deflect_timer_expire() 98 cs->next->prev = cs->prev; /* back link */ deflect_timer_expire() 191 cs->prev = NULL; cf_command() 295 ds->prev = table_tail; /* previous entry */ insertrule() 297 if (ds->prev) insertrule() 298 ds->prev->next = ds; /* last forward */ insertrule() 304 ds->prev = ds1->prev; /* prev entry */ insertrule() 305 ds1->prev = ds; /* backward chain old element */ insertrule() 306 if (!ds->prev) insertrule() 350 ds->next->prev = ds->prev; /* backward chain */ deleterule() 352 table_tail = ds->prev; /* end of chain */ deleterule() 354 if (ds->prev) deleterule() 355 ds->prev->next = ds->next; /* forward chain */ deleterule() 474 cs->prev = NULL; isdn_divert_icall() 518 cs->prev = NULL; isdn_divert_icall() 756 if (cs1->prev) prot_stat_callback() 757 cs1->prev->next = cs1->next; /* forward link */ prot_stat_callback() 761 cs1->next->prev = cs1->prev; /* back link */ prot_stat_callback() 812 if (cs1->prev) isdn_divert_stat_callback() 813 cs1->prev->next = cs1->next; /* forward link */ isdn_divert_stat_callback() 817 cs1->next->prev = cs1->prev; /* back link */ isdn_divert_stat_callback()
|
/linux-4.1.27/drivers/gpu/drm/ |
H A D | drm_lock.c | 189 unsigned int old, new, prev; drm_lock_take() local 202 prev = cmpxchg(lock, old, new); drm_lock_take() 203 } while (prev != old); drm_lock_take() 238 unsigned int old, new, prev; drm_lock_transfer() local 245 prev = cmpxchg(lock, old, new); drm_lock_transfer() 246 } while (prev != old); drm_lock_transfer() 263 unsigned int old, new, prev; drm_legacy_lock_free() local 278 prev = cmpxchg(lock, old, new); drm_legacy_lock_free() 279 } while (prev != old); drm_legacy_lock_free() 305 unsigned int old, new, prev; drm_notifier() local 317 prev = cmpxchg(&lock->lock, old, new); drm_notifier() 318 } while (prev != old); drm_notifier() 356 unsigned int old, prev; drm_legacy_idlelock_release() local 364 prev = cmpxchg(lock, old, DRM_KERNEL_CONTEXT); drm_legacy_idlelock_release() 365 } while (prev != old); drm_legacy_idlelock_release()
|
/linux-4.1.27/net/ceph/ |
H A D | pagelist.c | 11 struct page *page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_unmap_tail() 119 c->page_lru = pl->head.prev; ceph_pagelist_set_cursor() 137 while (pl->head.prev != c->page_lru) { ceph_pagelist_truncate() 138 page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_truncate() 145 page = list_entry(pl->head.prev, struct page, lru); ceph_pagelist_truncate()
|
/linux-4.1.27/net/netfilter/ |
H A D | xt_limit.c | 21 unsigned long prev; member in struct:xt_limit_priv 43 `prev' tracks the last packet hit: you gain one credit per jiffy. 76 priv->credit += (now - xchg(&priv->prev, now)) * CREDITS_PER_JIFFY; limit_mt() 123 priv->prev = jiffies; limit_mt_check() 144 compat_ulong_t prev; member in struct:compat_xt_rateinfo 151 /* To keep the full "prev" timestamp, the upper 32 bits are stored in the 159 .prev = cm->prev | (unsigned long)cm->master << 32, limit_mt_compat_from_user() 173 .prev = m->prev, limit_mt_compat_to_user() 177 .master = m->prev >> 32, limit_mt_compat_to_user()
|
/linux-4.1.27/arch/powerpc/boot/ |
H A D | ops.h | 47 void *(*find_node_by_prop_value)(const void *prev, 50 void *(*find_node_by_compatible)(const void *prev, 146 static inline void *find_node_by_prop_value(const void *prev, find_node_by_prop_value() argument 151 return dt_ops.find_node_by_prop_value(prev, propname, find_node_by_prop_value() 157 static inline void *find_node_by_prop_value_str(const void *prev, find_node_by_prop_value_str() argument 161 return find_node_by_prop_value(prev, propname, propval, find_node_by_prop_value_str() 165 static inline void *find_node_by_devtype(const void *prev, find_node_by_devtype() argument 168 return find_node_by_prop_value_str(prev, "device_type", type); find_node_by_devtype() 184 static inline void *find_node_by_compatible(const void *prev, find_node_by_compatible() argument 188 return dt_ops.find_node_by_compatible(prev, compat); find_node_by_compatible()
|
H A D | libfdt-wrapper.c | 131 static void *fdt_wrapper_find_node_by_prop_value(const void *prev, fdt_wrapper_find_node_by_prop_value() argument 136 int offset = fdt_node_offset_by_prop_value(fdt, devp_offset_find(prev), fdt_wrapper_find_node_by_prop_value() 141 static void *fdt_wrapper_find_node_by_compatible(const void *prev, fdt_wrapper_find_node_by_compatible() argument 144 int offset = fdt_node_offset_by_compatible(fdt, devp_offset_find(prev), fdt_wrapper_find_node_by_compatible()
|
/linux-4.1.27/sound/synth/ |
H A D | util_mem.c | 101 return __snd_util_memblk_new(hdr, units, p->prev); __snd_util_mem_alloc() 107 * the block is linked next to prev 111 struct list_head *prev) __snd_util_memblk_new() 120 if (prev == &hdr->block) __snd_util_memblk_new() 123 struct snd_util_memblk *p = get_memblk(prev); __snd_util_memblk_new() 127 list_add(&blk->list, prev); __snd_util_memblk_new() 110 __snd_util_memblk_new(struct snd_util_memhdr *hdr, unsigned int units, struct list_head *prev) __snd_util_memblk_new() argument
|
/linux-4.1.27/arch/metag/kernel/ |
H A D | process.c | 301 struct task_struct *__sched __switch_to(struct task_struct *prev, __switch_to() argument 307 to.Switch.pPara = prev; __switch_to() 310 if (prev->thread.user_flags & TBICTX_FPAC_BIT) { __switch_to() 311 struct pt_regs *regs = task_pt_regs(prev); __switch_to() 314 state.Sig.SaveMask = prev->thread.user_flags; __switch_to() 317 if (!prev->thread.fpu_context) __switch_to() 318 alloc_fpu_context(&prev->thread); __switch_to() 319 if (prev->thread.fpu_context) __switch_to() 320 __TBICtxFPUSave(state, prev->thread.fpu_context); __switch_to() 326 if (prev->thread.fpu_context) __switch_to() 327 prev->thread.fpu_context->needs_restore = true; __switch_to() 331 from = __TBISwitch(to, &prev->thread.kernel_context); __switch_to()
|
/linux-4.1.27/net/bridge/netfilter/ |
H A D | ebt_limit.c | 41 info->credit += (now - xchg(&info->prev, now)) * CREDITS_PER_JIFFY; ebt_limit_mt() 81 info->prev = jiffies; ebt_limit_mt_check() 96 compat_ulong_t prev; member in struct:ebt_compat_limit_info
|
/linux-4.1.27/drivers/media/pci/zoran/ |
H A D | videocodec.c | 172 struct attached_list *a, *prev; videocodec_detach() local 192 prev = NULL; videocodec_detach() 208 if (prev == NULL) { videocodec_detach() 213 prev->next = a->next; videocodec_detach() 223 prev = a; videocodec_detach() 271 struct codec_list *prev = NULL, *h = codeclist_top; videocodec_unregister() local 300 if (prev == NULL) { videocodec_unregister() 305 prev->next = h->next; videocodec_unregister() 312 prev = h; videocodec_unregister()
|
/linux-4.1.27/drivers/net/ethernet/sfc/ |
H A D | mdio_10g.c | 235 struct ethtool_cmd prev = { .cmd = ETHTOOL_GSET }; efx_mdio_set_settings() local 237 efx->phy_op->get_settings(efx, &prev); efx_mdio_set_settings() 239 if (ecmd->advertising == prev.advertising && efx_mdio_set_settings() 240 ethtool_cmd_speed(ecmd) == ethtool_cmd_speed(&prev) && efx_mdio_set_settings() 241 ecmd->duplex == prev.duplex && efx_mdio_set_settings() 242 ecmd->port == prev.port && efx_mdio_set_settings() 243 ecmd->autoneg == prev.autoneg) efx_mdio_set_settings() 247 if (prev.port != PORT_TP || ecmd->port != PORT_TP) efx_mdio_set_settings() 252 (ecmd->advertising | SUPPORTED_Autoneg) & ~prev.supported) efx_mdio_set_settings()
|
/linux-4.1.27/drivers/input/joystick/ |
H A D | joydump.c | 52 struct joydump *dump, *prev; /* one entry each */ joydump_connect() local 123 prev = dump; joydump_connect() 132 for (i = 1; i < t; i++, dump++, prev++) { joydump_connect() 134 i, dump->time - prev->time); joydump_connect()
|
/linux-4.1.27/drivers/usb/image/ |
H A D | microtek.h | 31 struct mts_desc *prev; member in struct:mts_desc
|
/linux-4.1.27/arch/arm/mach-omap2/ |
H A D | sdrc2xxx.c | 85 u32 prev = curr_perf_level; omap2xxx_sdrc_reprogram() local 89 return prev; omap2xxx_sdrc_reprogram() 96 return prev; omap2xxx_sdrc_reprogram() 113 return prev; omap2xxx_sdrc_reprogram()
|
/linux-4.1.27/fs/nfs/ |
H A D | pagelist.c | 269 * @prev - the previous request in page group, or NULL if @req is the first 273 nfs_page_group_init(struct nfs_page *req, struct nfs_page *prev) nfs_page_group_init() argument 276 WARN_ON_ONCE(prev == req); nfs_page_group_init() 278 if (!prev) { nfs_page_group_init() 284 WARN_ON_ONCE(prev->wb_this_page != prev->wb_head); nfs_page_group_init() 285 WARN_ON_ONCE(!test_bit(PG_HEADLOCK, &prev->wb_head->wb_flags)); nfs_page_group_init() 286 req->wb_head = prev->wb_head; nfs_page_group_init() 287 req->wb_this_page = prev->wb_this_page; nfs_page_group_init() 288 prev->wb_this_page = req; nfs_page_group_init() 297 if (test_bit(PG_INODE_REF, &prev->wb_head->wb_flags)) { nfs_page_group_init() 488 * @prev: previous request in desc, or NULL 495 struct nfs_page *prev, struct nfs_page *req) nfs_generic_pg_test() 922 * @prev: pointer to nfs_page 925 * The nfs_page structures 'prev' and 'req' are compared to ensure that the 931 static bool nfs_can_coalesce_requests(struct nfs_page *prev, nfs_can_coalesce_requests() argument 938 if (prev) { nfs_can_coalesce_requests() 939 if (!nfs_match_open_context(req->wb_context, prev->wb_context)) nfs_can_coalesce_requests() 946 prev->wb_lock_context)) nfs_can_coalesce_requests() 948 if (req_offset(req) != req_offset(prev) + prev->wb_bytes) nfs_can_coalesce_requests() 950 if (req->wb_page == prev->wb_page) { nfs_can_coalesce_requests() 951 if (req->wb_pgbase != prev->wb_pgbase + prev->wb_bytes) nfs_can_coalesce_requests() 955 prev->wb_pgbase + prev->wb_bytes != PAGE_CACHE_SIZE) nfs_can_coalesce_requests() 959 size = pgio->pg_ops->pg_test(pgio, prev, req); nfs_can_coalesce_requests() 979 struct nfs_page *prev = NULL; nfs_pageio_do_add_request() local 982 prev = nfs_list_entry(mirror->pg_list.prev); nfs_pageio_do_add_request() 988 if (!nfs_can_coalesce_requests(prev, req, desc)) nfs_pageio_do_add_request() 1273 struct nfs_page *prev; nfs_pageio_cond_complete() local 1279 prev = nfs_list_entry(mirror->pg_list.prev); nfs_pageio_cond_complete() 1280 if (index != prev->wb_index + 1) nfs_pageio_cond_complete() 494 nfs_generic_pg_test(struct nfs_pageio_descriptor *desc, struct nfs_page *prev, struct nfs_page *req) nfs_generic_pg_test() argument
|
/linux-4.1.27/fs/notify/dnotify/ |
H A D | dnotify.c | 93 struct dnotify_struct **prev; dnotify_handle_event() local 106 prev = &dn_mark->dn; dnotify_handle_event() 107 while ((dn = *prev) != NULL) { dnotify_handle_event() 109 prev = &dn->dn_next; dnotify_handle_event() 115 prev = &dn->dn_next; dnotify_handle_event() 117 *prev = dn->dn_next; dnotify_handle_event() 155 struct dnotify_struct **prev; dnotify_flush() local 170 prev = &dn_mark->dn; dnotify_flush() 171 while ((dn = *prev) != NULL) { dnotify_flush() 173 *prev = dn->dn_next; dnotify_flush() 178 prev = &dn->dn_next; dnotify_flush()
|
/linux-4.1.27/fs/hfs/ |
H A D | bnode.c | 144 be32_to_cpu(desc.next), be32_to_cpu(desc.prev), hfs_bnode_dump() 179 if (node->prev) { hfs_bnode_unlink() 180 tmp = hfs_bnode_find(tree, node->prev); hfs_bnode_unlink() 194 tmp->prev = node->prev; hfs_bnode_unlink() 195 cnid = cpu_to_be32(tmp->prev); hfs_bnode_unlink() 196 hfs_bnode_write(tmp, &cnid, offsetof(struct hfs_bnode_desc, prev), 4); hfs_bnode_unlink() 199 tree->leaf_tail = node->prev; hfs_bnode_unlink() 202 if (!node->prev && !node->next) { hfs_bnode_unlink() 342 node->prev = be32_to_cpu(desc->prev); hfs_bnode_find()
|
H A D | brec.c | 246 new_node->prev = node->this; hfs_bnode_split() 316 node_desc.prev = cpu_to_be32(new_node->prev); hfs_bnode_split() 332 next_node->prev = new_node->this; hfs_bnode_split() 334 node_desc.prev = cpu_to_be32(next_node->prev); hfs_bnode_split() 481 new_node->prev = 0; hfs_btree_inc_height() 485 node_desc.prev = cpu_to_be32(new_node->prev); hfs_btree_inc_height()
|
/linux-4.1.27/sound/core/seq/ |
H A D | seq_prioq.c | 149 struct snd_seq_event_cell *cur, *prev; snd_seq_prioq_cell_in() local 179 prev = NULL; /* previous cell */ snd_seq_prioq_cell_in() 194 prev = cur; snd_seq_prioq_cell_in() 204 if (prev != NULL) snd_seq_prioq_cell_in() 205 prev->next = cell; snd_seq_prioq_cell_in() 294 struct snd_seq_event_cell *prev = NULL; snd_seq_prioq_leave() local 307 prev->next = cell->next; snd_seq_prioq_leave() 329 prev = cell; snd_seq_prioq_leave() 404 struct snd_seq_event_cell *prev = NULL; snd_seq_prioq_remove_events() local 420 prev->next = cell->next; snd_seq_prioq_remove_events() 437 prev = cell; snd_seq_prioq_remove_events()
|
/linux-4.1.27/drivers/mtd/ |
H A D | redboot.c | 180 struct fis_list *new_fl, **prev; parse_redboot_partitions() local 207 prev = &fl; parse_redboot_partitions() 208 while(*prev && (*prev)->img->flash_base < new_fl->img->flash_base) parse_redboot_partitions() 209 prev = &(*prev)->next; parse_redboot_partitions() 210 new_fl->next = *prev; parse_redboot_partitions() 211 *prev = new_fl; parse_redboot_partitions()
|
/linux-4.1.27/drivers/dma/ |
H A D | txx9dmac.c | 174 return list_entry(dc->active_list.prev, txx9dmac_last_active() 186 desc = list_entry(desc->tx_list.prev, typeof(*desc), desc_node); txx9dmac_last_child() 435 struct txx9dmac_desc *prev = NULL; txx9dmac_dequeue() local 440 if (prev) { txx9dmac_dequeue() 441 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_dequeue() 443 prev->txd.phys, ddev->descsize, txx9dmac_dequeue() 446 prev = txx9dmac_last_child(desc); txx9dmac_dequeue() 723 struct txx9dmac_desc *prev; txx9dmac_prep_dma_memcpy() local 735 prev = first = NULL; txx9dmac_prep_dma_memcpy() 785 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_prep_dma_memcpy() 787 prev->txd.phys, ddev->descsize, txx9dmac_prep_dma_memcpy() 791 prev = desc; txx9dmac_prep_dma_memcpy() 796 txx9dmac_desc_set_INTENT(ddev, prev); txx9dmac_prep_dma_memcpy() 798 desc_write_CHAR(dc, prev, 0); txx9dmac_prep_dma_memcpy() 800 prev->txd.phys, ddev->descsize, txx9dmac_prep_dma_memcpy() 817 struct txx9dmac_desc *prev; txx9dmac_prep_slave_sg() local 832 prev = first = NULL; txx9dmac_prep_slave_sg() 879 desc_write_CHAR(dc, prev, desc->txd.phys); for_each_sg() 881 prev->txd.phys, for_each_sg() 886 prev = desc; for_each_sg() 891 txx9dmac_desc_set_INTENT(ddev, prev); 893 desc_write_CHAR(dc, prev, 0); 895 prev->txd.phys, ddev->descsize, 947 struct txx9dmac_desc *prev) txx9dmac_chain_dynamic() 953 prev = txx9dmac_last_child(prev); txx9dmac_chain_dynamic() 956 desc_write_CHAR(dc, prev, desc->txd.phys); txx9dmac_chain_dynamic() 958 prev->txd.phys, ddev->descsize, txx9dmac_chain_dynamic() 962 channel_read_CHAR(dc) == prev->txd.phys) txx9dmac_chain_dynamic() 981 struct txx9dmac_desc *prev = txx9dmac_last_active(dc); txx9dmac_issue_pending() local 983 if (!(prev->txd.flags & DMA_PREP_INTERRUPT) || txx9dmac_issue_pending() 985 txx9dmac_chain_dynamic(dc, prev); txx9dmac_issue_pending() 946 txx9dmac_chain_dynamic(struct txx9dmac_chan *dc, struct txx9dmac_desc *prev) txx9dmac_chain_dynamic() argument
|
H A D | img-mdc-dma.c | 294 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_dma_memcpy() local 313 if (prev) { mdc_prep_dma_memcpy() 314 prev->node_addr = curr_phys; mdc_prep_dma_memcpy() 315 prev->next_desc = curr; mdc_prep_dma_memcpy() 326 prev = curr; mdc_prep_dma_memcpy() 377 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_dma_cyclic() local 409 if (!prev) { mdc_prep_dma_cyclic() 413 prev->node_addr = curr_phys; mdc_prep_dma_cyclic() 414 prev->next_desc = curr; mdc_prep_dma_cyclic() 432 prev = curr; mdc_prep_dma_cyclic() 441 prev->node_addr = mdesc->list_phys; mdc_prep_dma_cyclic() 460 struct mdc_hw_list_desc *curr, *prev = NULL; mdc_prep_slave_sg() local 490 if (!prev) { for_each_sg() 494 prev->node_addr = curr_phys; for_each_sg() 495 prev->next_desc = curr; for_each_sg() 511 prev = curr; for_each_sg()
|
/linux-4.1.27/arch/c6x/kernel/ |
H A D | switch_to.S | 16 * void __switch_to(struct thread_info *prev, 22 || MV .L2X A4,B5 ; prev
|
/linux-4.1.27/arch/arc/kernel/ |
H A D | fpu.c | 33 void fpu_save_restore(struct task_struct *prev, struct task_struct *next) fpu_save_restore() argument 35 unsigned int *saveto = &prev->thread.fpu.aux_dpfp[0].l; fpu_save_restore()
|
H A D | ctx_sw.c | 26 unsigned int prev = (unsigned int)prev_task; __switch_to() local 105 /* last (ret value) = prev : although for ARC it mov r0, r0 */ __switch_to() 111 : "n"(KSP_WORD_OFF), "r"(next), "r"(prev) __switch_to()
|
/linux-4.1.27/kernel/trace/ |
H A D | trace_sched_switch.c | 19 probe_sched_switch(void *ignore, struct task_struct *prev, struct task_struct *next) probe_sched_switch() argument 24 tracing_record_cmdline(prev); probe_sched_switch()
|
H A D | trace_stat.h | 16 void *(*stat_next)(void *prev, int idx);
|
/linux-4.1.27/net/ieee802154/6lowpan/ |
H A D | reassembly.c | 49 struct sk_buff *prev, struct net_device *dev); 143 struct sk_buff *prev, *next; lowpan_frag_queue() local 176 prev = fq->q.fragments_tail; lowpan_frag_queue() 177 if (!prev || lowpan_cb(prev)->d_offset < lowpan_cb(skb)->d_offset) { lowpan_frag_queue() 181 prev = NULL; lowpan_frag_queue() 185 prev = next; lowpan_frag_queue() 193 if (prev) lowpan_frag_queue() 194 prev->next = skb; lowpan_frag_queue() 218 res = lowpan_frag_reasm(fq, prev, dev); lowpan_frag_queue() 237 static int lowpan_frag_reasm(struct lowpan_frag_queue *fq, struct sk_buff *prev, lowpan_frag_reasm() argument 246 if (prev) { lowpan_frag_reasm() 247 head = prev->next; lowpan_frag_reasm() 256 prev->next = fp; lowpan_frag_reasm()
|
/linux-4.1.27/arch/sparc/kernel/ |
H A D | sun4m_smp.c | 123 int *prev; smp4m_smp_done() local 127 prev = &first; for_each_online_cpu() 129 *prev = i; for_each_online_cpu() 130 prev = &cpu_data(i).next; for_each_online_cpu() 132 *prev = first;
|
H A D | sun4d_smp.c | 165 int *prev; smp4d_smp_done() local 169 prev = &first; for_each_online_cpu() 171 *prev = i; for_each_online_cpu() 172 prev = &cpu_data(i).next; for_each_online_cpu() 174 *prev = first;
|
/linux-4.1.27/fs/gfs2/ |
H A D | xattr.c | 77 struct gfs2_ea_header *prev, void *private); 82 struct gfs2_ea_header *ea, *prev = NULL; ea_foreach_i() local 88 for (ea = GFS2_EA_BH2FIRST(bh);; prev = ea, ea = GFS2_EA2NEXT(ea)) { ea_foreach_i() 97 error = ea_call(ip, bh, ea, prev, data); ea_foreach_i() 167 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_find_i() 182 el->el_prev = prev; ea_find_i() 215 * @prev: 227 struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() 297 if (prev && !leave) { ea_dealloc_unstuffed() 300 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_dealloc_unstuffed() 301 prev->ea_rec_len = cpu_to_be32(len); ea_dealloc_unstuffed() 304 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_dealloc_unstuffed() 327 struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() 339 error = ea_dealloc_unstuffed(ip, bh, ea, prev, (leave) ? &error : NULL); ea_remove_unstuffed() 366 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_list_i() 836 struct gfs2_ea_header *prev = el->el_prev; ea_set_remove_stuffed() local 841 if (!prev || !GFS2_EA_IS_STUFFED(ea)) { ea_set_remove_stuffed() 844 } else if (GFS2_EA2NEXT(prev) != ea) { ea_set_remove_stuffed() 845 prev = GFS2_EA2NEXT(prev); ea_set_remove_stuffed() 846 gfs2_assert_withdraw(GFS2_SB(&ip->i_inode), GFS2_EA2NEXT(prev) == ea); ea_set_remove_stuffed() 849 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_set_remove_stuffed() 850 prev->ea_rec_len = cpu_to_be32(len); ea_set_remove_stuffed() 853 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_set_remove_stuffed() 922 struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, ea_set_simple() 937 error = ea_remove_unstuffed(ip, bh, ea, prev, 1); ea_set_simple() 1089 struct gfs2_ea_header *prev = el->el_prev; ea_remove_stuffed() local 1099 if (prev) { ea_remove_stuffed() 1102 len = GFS2_EA_REC_LEN(prev) + GFS2_EA_REC_LEN(ea); ea_remove_stuffed() 1103 prev->ea_rec_len = cpu_to_be32(len); ea_remove_stuffed() 1106 prev->ea_flags |= GFS2_EAFLAG_LAST; ea_remove_stuffed() 166 ea_find_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_find_i() argument 225 ea_dealloc_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_dealloc_unstuffed() argument 325 ea_remove_unstuffed(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, int leave) ea_remove_unstuffed() argument 365 ea_list_i(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_list_i() argument 921 ea_set_simple(struct gfs2_inode *ip, struct buffer_head *bh, struct gfs2_ea_header *ea, struct gfs2_ea_header *prev, void *private) ea_set_simple() argument
|
/linux-4.1.27/drivers/usb/host/whci/ |
H A D | asl.c | 29 struct whc_qset **next, struct whc_qset **prev) qset_get_next_prev() 38 p = qset->list_node.prev; qset_get_next_prev() 40 p = p->prev; qset_get_next_prev() 43 *prev = container_of(p, struct whc_qset, list_node); qset_get_next_prev() 55 struct whc_qset *next, *prev; asl_qset_insert() local 60 qset_get_next_prev(whc, qset, &next, &prev); asl_qset_insert() 62 whc_qset_set_link_ptr(&prev->qh.link, qset->qset_dma); asl_qset_insert() 68 struct whc_qset *prev, *next; asl_qset_remove() local 70 qset_get_next_prev(whc, qset, &next, &prev); asl_qset_remove() 83 whc_qset_set_link_ptr(&prev->qh.link, next->qset_dma); asl_qset_remove() 28 qset_get_next_prev(struct whc *whc, struct whc_qset *qset, struct whc_qset **next, struct whc_qset **prev) qset_get_next_prev() argument
|
/linux-4.1.27/net/ipv4/ |
H A D | ip_fragment.c | 95 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, 320 struct sk_buff *prev, *next; ip_frag_queue() local 386 prev = qp->q.fragments_tail; ip_frag_queue() 387 if (!prev || FRAG_CB(prev)->offset < offset) { ip_frag_queue() 391 prev = NULL; ip_frag_queue() 395 prev = next; ip_frag_queue() 403 if (prev) { ip_frag_queue() 404 int i = (FRAG_CB(prev)->offset + prev->len) - offset; ip_frag_queue() 443 if (prev) ip_frag_queue() 444 prev->next = next; ip_frag_queue() 460 if (prev) ip_frag_queue() 461 prev->next = skb; ip_frag_queue() 486 err = ip_frag_reasm(qp, prev, dev); ip_frag_queue() 502 static int ip_frag_reasm(struct ipq *qp, struct sk_buff *prev, ip_frag_reasm() argument 522 if (prev) { ip_frag_reasm() 523 head = prev->next; ip_frag_reasm() 531 prev->next = fp; ip_frag_reasm()
|
/linux-4.1.27/drivers/iommu/ |
H A D | iova.c | 142 struct rb_node *prev, *curr = NULL; __alloc_and_insert_iova_range() local 151 prev = curr; __alloc_and_insert_iova_range() 168 prev = curr; __alloc_and_insert_iova_range() 190 /* If we have 'prev', it's a valid place to start the __alloc_and_insert_iova_range() 192 if (prev) __alloc_and_insert_iova_range() 193 entry = &prev; __alloc_and_insert_iova_range() 498 struct iova *prev = NULL, *next = NULL; split_and_remove_iova() local 502 prev = alloc_and_init_iova(iova->pfn_lo, pfn_lo - 1); split_and_remove_iova() 503 if (prev == NULL) split_and_remove_iova() 515 if (prev) { split_and_remove_iova() 516 iova_insert_rbtree(&iovad->rbroot, prev); split_and_remove_iova() 529 if (prev) split_and_remove_iova() 530 free_iova_mem(prev); split_and_remove_iova()
|
/linux-4.1.27/drivers/staging/lustre/lustre/fld/ |
H A D | fld_cache.c | 227 curr = cache->fci_lru.prev; fld_cache_shrink() 233 curr = curr->prev; fld_cache_shrink() 309 * prev range only. fixup will handle overlapping with next range. */ fld_cache_overlap_handle() 344 fld_cache_entry_add(cache, f_new, f_curr->fce_list.prev); fld_cache_overlap_handle() 387 struct list_head *prev = NULL; fld_cache_insert_nolock() local 410 prev = &f_curr->fce_list; list_for_each_entry_safe() 419 if (prev == NULL) 420 prev = head; 424 fld_cache_entry_add(cache, f_new, prev); 521 struct fld_cache_entry *prev = NULL; fld_cache_lookup() local 530 if (prev != NULL) list_for_each_entry() 531 *range = prev->fce_range; list_for_each_entry() 535 prev = flde; list_for_each_entry()
|
/linux-4.1.27/arch/cris/arch-v10/kernel/ |
H A D | fasttimer.c | 260 t->prev = NULL; start_one_shot_timer() 264 fast_timer_list->prev = t; start_one_shot_timer() 279 t->prev = tmp; start_one_shot_timer() 283 tmp->next->prev = t; start_one_shot_timer() 296 return (t->next != NULL) || (t->prev != NULL) || (t == fast_timer_list); fast_timer_pending() 301 struct fast_timer *next, *prev; detach_fast_timer() local 305 prev = t->prev; detach_fast_timer() 307 next->prev = prev; detach_fast_timer() 308 if (prev) detach_fast_timer() 309 prev->next = next; detach_fast_timer() 323 t->next = t->prev = NULL; del_fast_timer() 384 if (t->prev) timer1_handler() 386 t->prev->next = t->next; timer1_handler() 394 t->next->prev = t->prev; timer1_handler() 396 t->prev = NULL; timer1_handler()
|
/linux-4.1.27/arch/cris/arch-v32/kernel/ |
H A D | fasttimer.c | 238 t->prev = NULL; start_one_shot_timer() 241 fast_timer_list->prev = t; start_one_shot_timer() 253 t->prev = tmp; start_one_shot_timer() 257 tmp->next->prev = t; start_one_shot_timer() 270 return (t->next != NULL) || (t->prev != NULL) || (t == fast_timer_list); fast_timer_pending() 275 struct fast_timer *next, *prev; detach_fast_timer() local 279 prev = t->prev; detach_fast_timer() 281 next->prev = prev; detach_fast_timer() 282 if (prev) detach_fast_timer() 283 prev->next = next; detach_fast_timer() 297 t->next = t->prev = NULL; del_fast_timer() 373 if (t->prev) timer_trig_handler() 374 t->prev->next = t->next; timer_trig_handler() 378 t->next->prev = t->prev; timer_trig_handler() 379 t->prev = NULL; timer_trig_handler()
|
/linux-4.1.27/drivers/mmc/core/ |
H A D | sdio_cis.c | 231 struct sdio_func_tuple *this, **prev; sdio_read_cis() local 255 prev = &func->tuples; sdio_read_cis() 257 prev = &card->tuples; sdio_read_cis() 259 BUG_ON(*prev); sdio_read_cis() 311 *prev = this; sdio_read_cis() 312 prev = &this->next; sdio_read_cis() 341 *prev = card->tuples; sdio_read_cis()
|
/linux-4.1.27/arch/sh/kernel/cpu/sh5/ |
H A D | switchto.S | 24 r2 - prev 25 r3 - &prev->thread 30 r2 - last (=prev) : this just stays in r2 throughout 120 st.l r3, 0, r15 ! prev->thread.sp 121 st.l r3, 8, r1 ! prev->thread.kregs 122 st.l r3, 4, r9 ! prev->thread.pc 123 st.q r1, 0, r9 ! save prev->thread.pc into pt_regs->pc
|
/linux-4.1.27/arch/m68k/sun3/ |
H A D | sun3dvma.c | 110 struct hole *prev = NULL; refill() local 117 if(!prev) { refill() 118 prev = hole; refill() 122 if(hole->end == prev->start) { refill() 123 hole->size += prev->size; refill() 124 hole->end = prev->end; refill() 125 list_move(&(prev->list), &hole_cache); refill()
|
/linux-4.1.27/fs/autofs4/ |
H A D | expire.c | 76 static struct dentry *get_next_positive_subdir(struct dentry *prev, get_next_positive_subdir() argument 86 if (prev) get_next_positive_subdir() 87 next = prev->d_child.next; get_next_positive_subdir() 89 prev = dget_dlock(root); get_next_positive_subdir() 90 next = prev->d_subdirs.next; get_next_positive_subdir() 97 dput(prev); get_next_positive_subdir() 115 dput(prev); get_next_positive_subdir() 123 static struct dentry *get_next_positive_dentry(struct dentry *prev, get_next_positive_dentry() argument 130 if (prev == NULL) get_next_positive_dentry() 135 p = prev; get_next_positive_dentry() 146 dput(prev); get_next_positive_dentry() 178 dput(prev); get_next_positive_dentry()
|
/linux-4.1.27/kernel/gcov/ |
H A D | base.c | 131 struct gcov_info *prev = NULL; gcov_module_notifier() local 140 gcov_info_unlink(prev, info); gcov_module_notifier() 144 prev = info; gcov_module_notifier()
|
/linux-4.1.27/net/atm/ |
H A D | mpoa_caches.c | 112 entry->prev = NULL; in_cache_add_entry() 114 client->in_cache->prev = entry; in_cache_add_entry() 200 if (entry->prev != NULL) in_cache_remove_entry() 201 entry->prev->next = entry->next; in_cache_remove_entry() 205 entry->next->prev = entry->prev; in_cache_remove_entry() 435 if (entry->prev != NULL) eg_cache_remove_entry() 436 entry->prev->next = entry->next; eg_cache_remove_entry() 440 entry->next->prev = entry->prev; eg_cache_remove_entry() 475 entry->prev = NULL; eg_cache_add_entry() 477 client->eg_cache->prev = entry; eg_cache_add_entry()
|
H A D | mpoa_caches.h | 16 struct in_cache_entry *prev; member in struct:in_cache_entry 53 struct eg_cache_entry *prev; member in struct:eg_cache_entry
|
/linux-4.1.27/net/ipv6/ |
H A D | reassembly.c | 79 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, 212 struct sk_buff *prev, *next; ip6_frag_queue() local 289 prev = fq->q.fragments_tail; ip6_frag_queue() 290 if (!prev || FRAG6_CB(prev)->offset < offset) { ip6_frag_queue() 294 prev = NULL; ip6_frag_queue() 298 prev = next; ip6_frag_queue() 310 if (prev && ip6_frag_queue() 311 (FRAG6_CB(prev)->offset + prev->len) > offset) ip6_frag_queue() 324 if (prev) ip6_frag_queue() 325 prev->next = skb; ip6_frag_queue() 353 res = ip6_frag_reasm(fq, prev, dev); ip6_frag_queue() 379 static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, ip6_frag_reasm() argument 396 if (prev) { ip6_frag_reasm() 397 head = prev->next; ip6_frag_reasm() 406 prev->next = fp; ip6_frag_reasm()
|
/linux-4.1.27/drivers/md/bcache/ |
H A D | bset.c | 280 /* Space required for the prev pointers */ bset_prev_bytes() 293 kfree(t->prev); bch_btree_keys_free() 295 free_pages((unsigned long) t->prev, bch_btree_keys_free() 306 t->prev = NULL; bch_btree_keys_free() 330 t->prev = bset_prev_bytes(b) < PAGE_SIZE bch_btree_keys_alloc() 333 if (!t->prev) bch_btree_keys_alloc() 499 * bits we're going to store in bkey_float->mantissa. t->prev[j] stores the size 528 return (void *) (((uint64_t *) tree_to_bkey(t, j)) - t->prev[j]); tree_to_prev_bkey() 533 * maintain a full search tree, we just keep a simple lookup table in t->prev. 537 return cacheline_to_bkey(t, cacheline, t->prev[cacheline]); table_to_bkey() 596 t->prev = t[-1].prev + j; bset_alloc_tree() 613 t->prev[0] = bkey_to_cacheline_offset(t, 0, t->data->start); bch_bset_build_unwritten_tree() 637 struct bkey *prev = NULL, *k = t->data->start; bch_bset_build_written_tree() local 660 prev = k, k = bkey_next(k); bch_bset_build_written_tree() 662 t->prev[j] = bkey_u64s(prev); bch_bset_build_written_tree() 750 t->prev[j] += shift; bch_bset_fix_lookup_table() 752 if (t->prev[j] > 7) { bch_bset_fix_lookup_table() 758 t->prev[j] = bkey_to_cacheline_offset(t, j, k); bch_bset_fix_lookup_table() 771 t->prev[t->size] = bkey_to_cacheline_offset(t, t->size, k); bch_bset_fix_lookup_table() 824 struct bkey *m, *prev = NULL; bch_btree_insert_key() local 840 prev = m, m = bkey_next(m); bch_btree_insert_key() 842 /* prev is in the tree, if we merge we're done */ bch_btree_insert_key() 844 if (prev && bch_btree_insert_key() 845 bch_bkey_try_merge(b, prev, k)) bch_btree_insert_key()
|
/linux-4.1.27/drivers/media/pci/cx88/ |
H A D | cx88-vbi.c | 167 struct cx88_buffer *prev; buffer_queue() local 184 prev = list_entry(q->active.prev, struct cx88_buffer, list); buffer_queue() 187 prev->risc.jmp[1] = cpu_to_le32(buf->risc.dma); buffer_queue()
|
/linux-4.1.27/drivers/scsi/ |
H A D | scsi_module.c | 47 while ((l = l->prev) != &sht->legacy_hosts) init_this_scsi_driver()
|
/linux-4.1.27/arch/sh/kernel/ |
H A D | process_32.c | 181 __switch_to(struct task_struct *prev, struct task_struct *next) __switch_to() argument 189 unlazy_fpu(prev, task_pt_regs(prev)); __switch_to() 213 return prev; __switch_to()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/bios/ |
H A D | base.c | 59 u8 *prev = bios->data; nvbios_extend() local 61 bios->data = prev; nvbios_extend() 64 memcpy(bios->data, prev, bios->size); nvbios_extend() 66 kfree(prev); nvbios_extend()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/volt/ |
H A D | base.c | 90 int prev = nvkm_volt_get(volt); nvkm_volt_set_id() local 91 if (!condition || prev < 0 || nvkm_volt_set_id() 92 (condition < 0 && ret < prev) || nvkm_volt_set_id() 93 (condition > 0 && ret > prev)) { nvkm_volt_set_id()
|
/linux-4.1.27/arch/arm64/include/asm/ |
H A D | mmu_context.h | 192 switch_mm(struct mm_struct *prev, struct mm_struct *next, switch_mm() argument 206 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next)) || prev != next) switch_mm() 211 #define activate_mm(prev,next) switch_mm(prev, next, NULL)
|
/linux-4.1.27/tools/include/linux/ |
H A D | types.h | 64 struct list_head *next, *prev; member in struct:list_head
|
/linux-4.1.27/fs/hfsplus/ |
H A D | bnode.c | 308 be32_to_cpu(desc.next), be32_to_cpu(desc.prev), hfs_bnode_dump() 343 if (node->prev) { hfs_bnode_unlink() 344 tmp = hfs_bnode_find(tree, node->prev); hfs_bnode_unlink() 359 tmp->prev = node->prev; hfs_bnode_unlink() 360 cnid = cpu_to_be32(tmp->prev); hfs_bnode_unlink() 362 offsetof(struct hfs_bnode_desc, prev), 4); hfs_bnode_unlink() 365 tree->leaf_tail = node->prev; hfs_bnode_unlink() 368 if (!node->prev && !node->next) hfs_bnode_unlink() 510 node->prev = be32_to_cpu(desc->prev); hfs_bnode_find()
|
H A D | brec.c | 250 new_node->prev = node->this; hfs_bnode_split() 320 node_desc.prev = cpu_to_be32(new_node->prev); hfs_bnode_split() 336 next_node->prev = new_node->this; hfs_bnode_split() 338 node_desc.prev = cpu_to_be32(next_node->prev); hfs_bnode_split() 485 new_node->prev = 0; hfs_btree_inc_height() 489 node_desc.prev = cpu_to_be32(new_node->prev); hfs_btree_inc_height()
|
/linux-4.1.27/block/ |
H A D | noop-iosched.c | 47 if (rq->queuelist.prev == &nd->queue) noop_former_request() 49 return list_entry(rq->queuelist.prev, struct request, queuelist); noop_former_request()
|
/linux-4.1.27/net/core/ |
H A D | dst.c | 301 unsigned long prev, new; dst_cow_metrics_generic() local 306 prev = cmpxchg(&dst->_metrics, old, new); dst_cow_metrics_generic() 308 if (prev != old) { dst_cow_metrics_generic() 310 p = __DST_METRICS_PTR(prev); dst_cow_metrics_generic() 311 if (prev & DST_METRICS_READ_ONLY) dst_cow_metrics_generic() 322 unsigned long prev, new; __dst_destroy_metrics_generic() local 325 prev = cmpxchg(&dst->_metrics, old, new); __dst_destroy_metrics_generic() 326 if (prev == old) __dst_destroy_metrics_generic()
|
/linux-4.1.27/tools/perf/util/ |
H A D | cpumap.c | 49 int n, cpu, prev; cpu_map__read() local 53 prev = -1; cpu_map__read() 58 if (prev >= 0) { cpu_map__read() 59 int new_max = nr_cpus + cpu - prev - 1; cpu_map__read() 69 while (++prev < cpu) cpu_map__read() 70 tmp_cpus[nr_cpus++] = prev; cpu_map__read() 82 prev = cpu; cpu_map__read() 84 prev = -1; cpu_map__read()
|
/linux-4.1.27/drivers/acpi/acpica/ |
H A D | psloop.c | 322 union acpi_operand_object *prev; acpi_ps_link_module_code() local 329 prev = next = acpi_gbl_module_code_list; acpi_ps_link_module_code() 331 prev = next; acpi_ps_link_module_code() 339 if (!prev || acpi_ps_link_module_code() 340 ((prev->method.aml_start + prev->method.aml_length) != aml_start)) { acpi_ps_link_module_code() 367 if (!prev) { acpi_ps_link_module_code() 370 prev->method.mutex = method_obj; acpi_ps_link_module_code() 373 prev->method.aml_length += aml_length; acpi_ps_link_module_code()
|
H A D | psparse.c | 136 union acpi_parse_object *prev; acpi_ps_complete_this_op() local 161 prev = op->common.parent->common.value.arg; acpi_ps_complete_this_op() 162 if (!prev) { acpi_ps_complete_this_op() 251 if (prev == op) { acpi_ps_complete_this_op() 272 while (prev) { acpi_ps_complete_this_op() 276 next = prev->common.next; acpi_ps_complete_this_op() 285 prev->common.next = acpi_ps_complete_this_op() 291 prev->common.next = acpi_ps_complete_this_op() 296 prev = next; acpi_ps_complete_this_op()
|
H A D | utaddress.c | 141 struct acpi_address_range *prev; acpi_ut_remove_address_range() local 152 range_info = prev = acpi_gbl_address_range_list[space_id]; acpi_ut_remove_address_range() 155 if (range_info == prev) { /* Found at list head */ acpi_ut_remove_address_range() 159 prev->next = range_info->next; acpi_ut_remove_address_range() 175 prev = range_info; acpi_ut_remove_address_range()
|
/linux-4.1.27/drivers/ata/ |
H A D | pata_sc1200.c | 169 struct ata_device *prev = ap->private_data; sc1200_qc_issue() local 172 if (ata_dma_enabled(adev) && adev != prev && prev != NULL) { sc1200_qc_issue() 174 if ((ata_using_udma(adev) && !ata_using_udma(prev)) || sc1200_qc_issue() 175 (ata_using_udma(prev) && !ata_using_udma(adev))) sc1200_qc_issue()
|
/linux-4.1.27/drivers/gpu/drm/savage/ |
H A D | savage_bci.c | 216 dev_priv->head.prev = NULL; savage_freelist_init() 220 dev_priv->tail.prev = &dev_priv->head; savage_freelist_init() 231 entry->prev = &dev_priv->head; savage_freelist_init() 232 dev_priv->head.next->prev = entry; savage_freelist_init() 242 drm_savage_buf_priv_t *tail = dev_priv->tail.prev; savage_freelist_get() 261 drm_savage_buf_priv_t *prev = tail->prev; savage_freelist_get() local 262 prev->next = next; savage_freelist_get() 263 next->prev = prev; savage_freelist_get() 264 tail->next = tail->prev = NULL; savage_freelist_get() 275 drm_savage_buf_priv_t *entry = buf->dev_private, *prev, *next; savage_freelist_put() local 279 if (entry->next != NULL || entry->prev != NULL) { savage_freelist_put() 284 prev = &dev_priv->head; savage_freelist_put() 285 next = prev->next; savage_freelist_put() 286 prev->next = entry; savage_freelist_put() 287 next->prev = entry; savage_freelist_put() 288 entry->prev = prev; savage_freelist_put() 1063 buf_priv->next == NULL && buf_priv->prev == NULL) { savage_reclaim_buffers()
|
/linux-4.1.27/drivers/gpu/host1x/ |
H A D | intr.c | 71 struct host1x_waitlist *waiter, *next, *prev; remove_completed_waiters() local 82 prev = list_entry(dest->prev, list_for_each_entry_safe() 84 if (prev->data == waiter->data) { list_for_each_entry_safe() 85 prev->count++; list_for_each_entry_safe()
|