/linux/drivers/net/ethernet/huawei/hinic/ |
A D | hinic_hw_wq.c | 140 &wqs->page_paddr[page_idx], in wqs_allocate_page() 156 wqs->page_vaddr[page_idx], in wqs_free_page() 251 *page_idx = wqs->free_blocks[pos].page_idx; in wqs_next_block() 271 wqs->free_blocks[pos].page_idx = page_idx; in wqs_return_block() 283 for (page_idx = 0; page_idx < wqs->num_pages; page_idx++) { in init_wqs_blocks_arr() 285 wqs->free_blocks[pos].page_idx = page_idx; in init_wqs_blocks_arr() 310 int err, i, page_idx; in hinic_wqs_alloc() local 327 for (page_idx = 0; page_idx < wqs->num_pages; page_idx++) { in hinic_wqs_alloc() 362 int page_idx; in hinic_wqs_free() local 366 for (page_idx = 0; page_idx < wqs->num_pages; page_idx++) in hinic_wqs_free() [all …]
|
A D | hinic_hw_wq.h | 18 int page_idx; member 25 int page_idx; member
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | dm.c | 24 u64 page_idx = 0; in mlx5_cmd_alloc_memic() local 48 page_idx, in mlx5_cmd_alloc_memic() 53 page_idx, num_pages); in mlx5_cmd_alloc_memic() 67 page_idx, num_pages); in mlx5_cmd_alloc_memic() 71 page_idx++; in mlx5_cmd_alloc_memic() 177 u16 page_idx; in copy_op_to_user() local 183 &page_idx, sizeof(page_idx)); in copy_op_to_user() 281 u16 page_idx; in handle_alloc_dm_memic() local 319 &page_idx, sizeof(page_idx)); in handle_alloc_dm_memic() 491 u16 page_idx; in UVERBS_HANDLER() local [all …]
|
A D | main.c | 2135 clear_bit(mentry->page_idx, var_table->bitmap); in mlx5_ib_mmap_free() 3369 u32 page_idx; in alloc_var_entry() local 3378 page_idx = find_first_zero_bit(var_table->bitmap, in alloc_var_entry() 3380 if (page_idx >= var_table->num_var_hw_entries) { in alloc_var_entry() 3386 set_bit(page_idx, var_table->bitmap); in alloc_var_entry() 3390 (page_idx * var_table->stride_size); in alloc_var_entry() 3391 entry->page_idx = page_idx; in alloc_var_entry() 3403 clear_bit(page_idx, var_table->bitmap); in alloc_var_entry() 3440 &entry->page_idx, sizeof(entry->page_idx)); in UVERBS_HANDLER() 3503 entry->page_idx = uar_index; in alloc_uar_entry() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | icm.h | 75 int page_idx; member 103 iter->page_idx = 0; in mlx4_icm_first() 113 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next() 121 iter->page_idx = 0; in mlx4_icm_next() 128 return iter->chunk->buf[iter->page_idx].dma_addr; in mlx4_icm_addr() 130 return sg_dma_address(&iter->chunk->sg[iter->page_idx]); in mlx4_icm_addr() 136 return iter->chunk->buf[iter->page_idx].size; in mlx4_icm_size() 138 return sg_dma_len(&iter->chunk->sg[iter->page_idx]); in mlx4_icm_size()
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_memfree.h | 77 int page_idx; member 106 iter->page_idx = 0; in mthca_icm_first() 116 if (++iter->page_idx >= iter->chunk->nsg) { in mthca_icm_next() 124 iter->page_idx = 0; in mthca_icm_next() 130 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mthca_icm_addr() 135 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mthca_icm_size()
|
/linux/drivers/infiniband/hw/hns/ |
A D | hns_roce_hem.h | 91 int page_idx; member 144 iter->page_idx = 0; in hns_roce_hem_first() 154 if (++iter->page_idx >= iter->chunk->nsg) { in hns_roce_hem_next() 162 iter->page_idx = 0; in hns_roce_hem_next() 168 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in hns_roce_hem_addr()
|
/linux/drivers/infiniband/sw/siw/ |
A D | siw_mem.h | 60 unsigned int page_idx = (addr - umem->fp_addr) >> PAGE_SHIFT, in siw_get_upage() local 61 chunk_idx = page_idx >> CHUNK_SHIFT, in siw_get_upage() 62 page_in_chunk = page_idx & ~CHUNK_MASK; in siw_get_upage() 64 if (likely(page_idx < umem->num_pages)) in siw_get_upage()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
A D | rx.c | 30 u32 page_idx) in mlx5e_xsk_skb_from_cqe_mpwrq_linear() argument 32 struct xdp_buff *xdp = wi->umr.dma_info[page_idx].xsk; in mlx5e_xsk_skb_from_cqe_mpwrq_linear() 70 __set_bit(page_idx, wi->xdp_xmit_bitmap); /* non-atomic */ in mlx5e_xsk_skb_from_cqe_mpwrq_linear()
|
A D | rx.h | 16 u32 page_idx);
|
/linux/sound/soc/amd/acp/ |
A D | acp-platform.c | 141 u16 page_idx; in config_acp_dma() local 145 for (page_idx = 0; page_idx < num_pages; page_idx++) { in config_acp_dma()
|
/linux/sound/soc/amd/yc/ |
A D | acp6x-pdm-dma.c | 149 u16 page_idx; in acp6x_config_dma() local 161 for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) { in acp6x_config_dma()
|
/linux/sound/soc/amd/renoir/ |
A D | acp3x-pdm-dma.c | 172 u16 page_idx; in config_acp_dma() local 185 for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) { in config_acp_dma()
|
/linux/fs/f2fs/ |
A D | file.c | 3414 pgoff_t page_idx = 0, last_idx; in f2fs_release_compress_blocks() local 3464 while (page_idx < last_idx) { in f2fs_release_compress_blocks() 3473 page_idx); in f2fs_release_compress_blocks() 3491 page_idx += count; in f2fs_release_compress_blocks() 3617 while (page_idx < last_idx) { in f2fs_reserve_compress_blocks() 3626 page_idx); in f2fs_reserve_compress_blocks() 3644 page_idx += count; in f2fs_reserve_compress_blocks() 4019 count = last_idx - page_idx; in f2fs_ioc_decompress_file() 4031 page_idx += len; in f2fs_ioc_decompress_file() 4090 count = last_idx - page_idx; in f2fs_ioc_compress_file() [all …]
|
/linux/sound/soc/amd/raven/ |
A D | acp3x-pcm-dma.c | 109 u16 page_idx; in config_acp3x_dma() local 141 for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) { in config_acp3x_dma()
|
/linux/sound/soc/amd/vangogh/ |
A D | acp5x-pcm-dma.c | 105 u16 page_idx; in config_acp5x_dma() local 136 for (page_idx = 0; page_idx < rtd->num_pages; page_idx++) { in config_acp5x_dma()
|
/linux/mm/ |
A D | percpu-vm.c | 14 unsigned int cpu, int page_idx) in pcpu_chunk_page() argument 19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page()
|
A D | percpu.c | 269 static int __maybe_unused pcpu_page_idx(unsigned int cpu, int page_idx) in pcpu_page_idx() argument 271 return pcpu_unit_map[cpu] * pcpu_unit_pages + page_idx; in pcpu_page_idx() 274 static unsigned long pcpu_unit_page_offset(unsigned int cpu, int page_idx) in pcpu_unit_page_offset() argument 276 return pcpu_unit_offsets[cpu] + (page_idx << PAGE_SHIFT); in pcpu_unit_page_offset() 280 unsigned int cpu, int page_idx) in pcpu_chunk_addr() argument 283 pcpu_unit_page_offset(cpu, page_idx); in pcpu_chunk_addr()
|
/linux/drivers/mmc/host/ |
A D | usdhi6rol0.c | 181 int page_idx; /* page index within an SG segment */ member 419 (host->page_idx << PAGE_SHIFT) + data->blksz - blk_head) in usdhi6_sg_unmap() 443 host->page_idx++; in usdhi6_sg_advance() 453 host->page_idx++; in usdhi6_sg_advance() 463 done = (host->page_idx << PAGE_SHIFT) + host->offset; in usdhi6_sg_advance() 489 host->page_idx = 0; in usdhi6_sg_advance() 506 host->pg.page = nth_page(sg_page(host->sg), host->page_idx); in usdhi6_sg_advance() 1011 host->page_idx = 0; in usdhi6_rq_start() 1722 data->flags & MMC_DATA_READ ? 'R' : 'W', host->page_idx, in usdhi6_timeout_work()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | en_rx.c | 62 u16 cqe_bcnt, u32 head_offset, u32 page_idx); 65 u16 cqe_bcnt, u32 head_offset, u32 page_idx); 1712 u32 page_idx = wqe_offset >> PAGE_SHIFT; in mlx5e_handle_rx_cqe_mpwrq_rep() local 1739 rq, wi, cqe_bcnt, head_offset, page_idx); in mlx5e_handle_rx_cqe_mpwrq_rep() 1789 u16 cqe_bcnt, u32 head_offset, u32 page_idx) in mlx5e_skb_from_cqe_mpwrq_nonlinear() argument 1792 struct mlx5e_dma_info *di = &wi->umr.dma_info[page_idx]; in mlx5e_skb_from_cqe_mpwrq_nonlinear() 1824 u16 cqe_bcnt, u32 head_offset, u32 page_idx) in mlx5e_skb_from_cqe_mpwrq_linear() argument 1976 u32 page_idx = wqe_offset >> PAGE_SHIFT; in mlx5e_handle_rx_cqe_mpwrq_shampo() local 2025 di = &wi->umr.dma_info[page_idx]; in mlx5e_handle_rx_cqe_mpwrq_shampo() 2051 u32 page_idx = wqe_offset >> PAGE_SHIFT; in mlx5e_handle_rx_cqe_mpwrq() local [all …]
|
/linux/fs/ubifs/ |
A D | file.c | 724 int err, page_idx, page_cnt, ret = 0, n = 0; in ubifs_do_bulk_read() local 783 for (page_idx = 1; page_idx < page_cnt; page_idx++) { in ubifs_do_bulk_read() 784 pgoff_t page_offset = offset + page_idx; in ubifs_do_bulk_read() 802 ui->last_page_read = offset + page_idx - 1; in ubifs_do_bulk_read()
|
/linux/sound/soc/amd/ |
A D | acp-pcm-dma.c | 301 u16 page_idx; in acp_pte_config() local 307 for (page_idx = 0; page_idx < (num_of_pages); page_idx++) { in acp_pte_config() 309 acp_reg_write((offset + (page_idx * 8)), in acp_pte_config() 318 acp_reg_write((offset + (page_idx * 8) + 4), in acp_pte_config()
|
/linux/drivers/input/touchscreen/ |
A D | raydium_i2c_ts.c | 646 u16 page_idx, const void *data, size_t len) in raydium_i2c_fw_write_page() argument 657 buf[BL_PAGE_STR] = page_idx ? 0xff : 0; in raydium_i2c_fw_write_page() 671 page_idx, i, error); in raydium_i2c_fw_write_page()
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_ttm.c | 845 uint64_t page_idx = 1; in amdgpu_ttm_gart_bind() local 847 r = amdgpu_gart_bind(adev, gtt->offset, page_idx, in amdgpu_ttm_gart_bind() 860 gtt->offset + (page_idx << PAGE_SHIFT), in amdgpu_ttm_gart_bind() 861 ttm->num_pages - page_idx, in amdgpu_ttm_gart_bind() 862 &(gtt->ttm.dma_address[page_idx]), flags); in amdgpu_ttm_gart_bind()
|
/linux/fs/ntfs3/ |
A D | fsntfs.c | 1471 u32 add, off, page_idx; in ntfs_bio_pages() local 1492 page_idx = 0; in ntfs_bio_pages() 1499 new = ntfs_alloc_bio(nr_pages - page_idx); in ntfs_bio_pages() 1526 page_idx += 1; in ntfs_bio_pages() 1527 if (WARN_ON(page_idx >= nr_pages)) { in ntfs_bio_pages() 1531 page = pages[page_idx]; in ntfs_bio_pages()
|