Lines Matching refs:bcnt
187 u32 key, u64 io_virt, size_t bcnt, in pagefault_single_data_segment() argument
216 (bcnt - pfault->mpfault.bytes_committed); in pagefault_single_data_segment()
238 bcnt -= pfault->mpfault.bytes_committed; in pagefault_single_data_segment()
244 npages = ib_umem_odp_map_dma_pages(mr->umem, io_virt, bcnt, in pagefault_single_data_segment()
273 *bytes_mapped += min_t(u32, new_mappings, bcnt); in pagefault_single_data_segment()
326 size_t bcnt; in pagefault_data_segments() local
345 bcnt = byte_count & ~MLX5_INLINE_SEG; in pagefault_data_segments()
348 bcnt = bcnt & MLX5_WQE_INLINE_SEG_BYTE_COUNT_MASK; in pagefault_data_segments()
349 wqe += ALIGN(sizeof(struct mlx5_wqe_inline_seg) + bcnt, in pagefault_data_segments()
356 if (receive_queue && bcnt == 0 && key == MLX5_INVALID_LKEY && in pagefault_data_segments()
361 *total_wqe_bytes += bcnt - min_t(size_t, bcnt, in pagefault_data_segments()
366 if (bcnt == 0) in pagefault_data_segments()
367 bcnt = 1U << 31; in pagefault_data_segments()
369 if (inline_segment || bcnt <= pfault->mpfault.bytes_committed) { in pagefault_data_segments()
371 min_t(size_t, bcnt, in pagefault_data_segments()
377 bcnt, bytes_mapped); in pagefault_data_segments()