Lines Matching refs:bcnt

552 			     u64 user_va, size_t bcnt, u32 *bytes_mapped,  in pagefault_real_mr()  argument
572 np = ib_umem_odp_map_dma_and_lock(odp, user_va, bcnt, access_mask, fault); in pagefault_real_mr()
594 *bytes_mapped += min_t(u32, new_mappings, bcnt); in pagefault_real_mr()
605 size_t bcnt, u32 *bytes_mapped, u32 flags) in pagefault_implicit_mr() argument
607 unsigned long end_idx = (user_va + bcnt - 1) >> MLX5_IMR_MTT_SHIFT; in pagefault_implicit_mr()
615 mlx5_imr_ksm_entries * MLX5_IMR_MTT_SIZE - user_va < bcnt)) in pagefault_implicit_mr()
619 while (bcnt) { in pagefault_implicit_mr()
642 len = min_t(u64, user_va + bcnt, ib_umem_end(umem_odp)) - in pagefault_implicit_mr()
653 bcnt -= len; in pagefault_implicit_mr()
689 static int pagefault_dmabuf_mr(struct mlx5_ib_mr *mr, size_t bcnt, in pagefault_dmabuf_mr() argument
722 *bytes_mapped += bcnt; in pagefault_dmabuf_mr()
736 static int pagefault_mr(struct mlx5_ib_mr *mr, u64 io_virt, size_t bcnt, in pagefault_mr() argument
745 return pagefault_dmabuf_mr(mr, bcnt, bytes_mapped, flags); in pagefault_mr()
754 ib_umem_end(odp) - user_va < bcnt)) in pagefault_mr()
756 return pagefault_real_mr(mr, odp, user_va, bcnt, bytes_mapped, in pagefault_mr()
759 return pagefault_implicit_mr(mr, odp, io_virt, bcnt, bytes_mapped, in pagefault_mr()
787 size_t bcnt; member
813 u64 io_virt, size_t bcnt, in pagefault_single_data_segment() argument
826 bcnt -= *bytes_committed; in pagefault_single_data_segment()
838 *bytes_mapped += bcnt; in pagefault_single_data_segment()
860 ret = pagefault_mr(mr, io_virt, bcnt, bytes_mapped, 0); in pagefault_single_data_segment()
901 for (i = 0; bcnt && i < mmkey->ndescs; i++, pklm++) { in pagefault_single_data_segment()
915 frame->bcnt = min_t(size_t, bcnt, in pagefault_single_data_segment()
921 bcnt -= frame->bcnt; in pagefault_single_data_segment()
938 bcnt = frame->bcnt; in pagefault_single_data_segment()
989 size_t bcnt; in pagefault_data_segments() local
1004 bcnt = byte_count & ~MLX5_INLINE_SEG; in pagefault_data_segments()
1007 bcnt = bcnt & MLX5_WQE_INLINE_SEG_BYTE_COUNT_MASK; in pagefault_data_segments()
1008 wqe += ALIGN(sizeof(struct mlx5_wqe_inline_seg) + bcnt, in pagefault_data_segments()
1015 if (receive_queue && bcnt == 0 && key == MLX5_INVALID_LKEY && in pagefault_data_segments()
1020 *total_wqe_bytes += bcnt - min_t(size_t, bcnt, in pagefault_data_segments()
1025 if (bcnt == 0) in pagefault_data_segments()
1026 bcnt = 1U << 31; in pagefault_data_segments()
1028 if (inline_segment || bcnt <= pfault->bytes_committed) { in pagefault_data_segments()
1030 min_t(size_t, bcnt, in pagefault_data_segments()
1036 io_virt, bcnt, in pagefault_data_segments()