Lines Matching refs:v

84 #define iterate_all_kinds(i, n, v, I, B, K) {			\  argument
88 struct bio_vec v; \
89 iterate_bvec(i, n, v, bvec, skip, (B)) \
92 struct kvec v; \
93 iterate_kvec(i, n, v, kvec, skip, (K)) \
96 struct iovec v; \
97 iterate_iovec(i, n, v, iov, skip, (I)) \
101 #define iterate_and_advance(i, n, v, I, B, K) { \ argument
105 struct bio_vec v; \
106 iterate_bvec(i, n, v, bvec, skip, (B)) \
115 struct kvec v; \
116 iterate_kvec(i, n, v, kvec, skip, (K)) \
125 struct iovec v; \
126 iterate_iovec(i, n, v, iov, skip, (I)) \
332 struct iovec v; in iov_iter_fault_in_multipages_readable() local
335 iterate_iovec(i, bytes, v, iov, skip, ({ in iov_iter_fault_in_multipages_readable()
336 err = fault_in_multipages_readable(v.iov_base, in iov_iter_fault_in_multipages_readable()
337 v.iov_len); in iov_iter_fault_in_multipages_readable()
395 iterate_and_advance(i, bytes, v, in copy_to_iter()
396 __copy_to_user(v.iov_base, (from += v.iov_len) - v.iov_len, in copy_to_iter()
397 v.iov_len), in copy_to_iter()
398 memcpy_to_page(v.bv_page, v.bv_offset, in copy_to_iter()
399 (from += v.bv_len) - v.bv_len, v.bv_len), in copy_to_iter()
400 memcpy(v.iov_base, (from += v.iov_len) - v.iov_len, v.iov_len) in copy_to_iter()
416 iterate_and_advance(i, bytes, v, in copy_from_iter()
417 __copy_from_user((to += v.iov_len) - v.iov_len, v.iov_base, in copy_from_iter()
418 v.iov_len), in copy_from_iter()
419 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in copy_from_iter()
420 v.bv_offset, v.bv_len), in copy_from_iter()
421 memcpy((to += v.iov_len) - v.iov_len, v.iov_base, v.iov_len) in copy_from_iter()
437 iterate_and_advance(i, bytes, v, in copy_from_iter_nocache()
438 __copy_from_user_nocache((to += v.iov_len) - v.iov_len, in copy_from_iter_nocache()
439 v.iov_base, v.iov_len), in copy_from_iter_nocache()
440 memcpy_from_page((to += v.bv_len) - v.bv_len, v.bv_page, in copy_from_iter_nocache()
441 v.bv_offset, v.bv_len), in copy_from_iter_nocache()
442 memcpy((to += v.iov_len) - v.iov_len, v.iov_base, v.iov_len) in copy_from_iter_nocache()
483 iterate_and_advance(i, bytes, v, in iov_iter_zero()
484 __clear_user(v.iov_base, v.iov_len), in iov_iter_zero()
485 memzero_page(v.bv_page, v.bv_offset, v.bv_len), in iov_iter_zero()
486 memset(v.iov_base, 0, v.iov_len) in iov_iter_zero()
497 iterate_all_kinds(i, bytes, v, in iov_iter_copy_from_user_atomic()
498 __copy_from_user_inatomic((p += v.iov_len) - v.iov_len, in iov_iter_copy_from_user_atomic()
499 v.iov_base, v.iov_len), in iov_iter_copy_from_user_atomic()
500 memcpy_from_page((p += v.bv_len) - v.bv_len, v.bv_page, in iov_iter_copy_from_user_atomic()
501 v.bv_offset, v.bv_len), in iov_iter_copy_from_user_atomic()
502 memcpy((p += v.iov_len) - v.iov_len, v.iov_base, v.iov_len) in iov_iter_copy_from_user_atomic()
511 iterate_and_advance(i, size, v, 0, 0, 0) in iov_iter_advance()
563 iterate_all_kinds(i, size, v, in iov_iter_alignment()
564 (res |= (unsigned long)v.iov_base | v.iov_len, 0), in iov_iter_alignment()
565 res |= v.bv_offset | v.bv_len, in iov_iter_alignment()
566 res |= (unsigned long)v.iov_base | v.iov_len in iov_iter_alignment()
582 iterate_all_kinds(i, maxsize, v, ({ in iov_iter_get_pages()
583 unsigned long addr = (unsigned long)v.iov_base; in iov_iter_get_pages()
584 size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); in iov_iter_get_pages()
598 *start = v.bv_offset; in iov_iter_get_pages()
599 get_page(*pages = v.bv_page); in iov_iter_get_pages()
600 return v.bv_len; in iov_iter_get_pages()
629 iterate_all_kinds(i, maxsize, v, ({ in iov_iter_get_pages_alloc()
630 unsigned long addr = (unsigned long)v.iov_base; in iov_iter_get_pages_alloc()
631 size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); in iov_iter_get_pages_alloc()
649 *start = v.bv_offset; in iov_iter_get_pages_alloc()
653 get_page(*p = v.bv_page); in iov_iter_get_pages_alloc()
654 return v.bv_len; in iov_iter_get_pages_alloc()
676 iterate_and_advance(i, bytes, v, ({ in csum_and_copy_from_iter()
678 next = csum_and_copy_from_user(v.iov_base, in csum_and_copy_from_iter()
679 (to += v.iov_len) - v.iov_len, in csum_and_copy_from_iter()
680 v.iov_len, 0, &err); in csum_and_copy_from_iter()
683 off += v.iov_len; in csum_and_copy_from_iter()
685 err ? v.iov_len : 0; in csum_and_copy_from_iter()
687 char *p = kmap_atomic(v.bv_page); in csum_and_copy_from_iter()
688 next = csum_partial_copy_nocheck(p + v.bv_offset, in csum_and_copy_from_iter()
689 (to += v.bv_len) - v.bv_len, in csum_and_copy_from_iter()
690 v.bv_len, 0); in csum_and_copy_from_iter()
693 off += v.bv_len; in csum_and_copy_from_iter()
695 next = csum_partial_copy_nocheck(v.iov_base, in csum_and_copy_from_iter()
696 (to += v.iov_len) - v.iov_len, in csum_and_copy_from_iter()
697 v.iov_len, 0); in csum_and_copy_from_iter()
699 off += v.iov_len; in csum_and_copy_from_iter()
720 iterate_and_advance(i, bytes, v, ({ in csum_and_copy_to_iter()
722 next = csum_and_copy_to_user((from += v.iov_len) - v.iov_len, in csum_and_copy_to_iter()
723 v.iov_base, in csum_and_copy_to_iter()
724 v.iov_len, 0, &err); in csum_and_copy_to_iter()
727 off += v.iov_len; in csum_and_copy_to_iter()
729 err ? v.iov_len : 0; in csum_and_copy_to_iter()
731 char *p = kmap_atomic(v.bv_page); in csum_and_copy_to_iter()
732 next = csum_partial_copy_nocheck((from += v.bv_len) - v.bv_len, in csum_and_copy_to_iter()
733 p + v.bv_offset, in csum_and_copy_to_iter()
734 v.bv_len, 0); in csum_and_copy_to_iter()
737 off += v.bv_len; in csum_and_copy_to_iter()
739 next = csum_partial_copy_nocheck((from += v.iov_len) - v.iov_len, in csum_and_copy_to_iter()
740 v.iov_base, in csum_and_copy_to_iter()
741 v.iov_len, 0); in csum_and_copy_to_iter()
743 off += v.iov_len; in csum_and_copy_to_iter()
759 iterate_all_kinds(i, size, v, ({ in iov_iter_npages()
760 unsigned long p = (unsigned long)v.iov_base; in iov_iter_npages()
761 npages += DIV_ROUND_UP(p + v.iov_len, PAGE_SIZE) in iov_iter_npages()
770 unsigned long p = (unsigned long)v.iov_base; in iov_iter_npages()
771 npages += DIV_ROUND_UP(p + v.iov_len, PAGE_SIZE) in iov_iter_npages()