/linux/drivers/dma-buf/heaps/ |
A D | cma_heap.c | 38 pgoff_t pagecount; member 62 buffer->pagecount, 0, in cma_heap_attach() 63 buffer->pagecount << PAGE_SHIFT, in cma_heap_attach() 166 if (vmf->pgoff > buffer->pagecount) in cma_heap_vm_fault() 257 cma_release(cma_heap->cma, buffer->cma_pages, buffer->pagecount); in cma_heap_dma_buf_release() 283 pgoff_t pagecount = size >> PAGE_SHIFT; in cma_heap_allocate() local 301 cma_pages = cma_alloc(cma_heap->cma, pagecount, align, false); in cma_heap_allocate() 307 unsigned long nr_clear_pages = pagecount; in cma_heap_allocate() 334 for (pg = 0; pg < pagecount; pg++) in cma_heap_allocate() 339 buffer->pagecount = pagecount; in cma_heap_allocate() [all …]
|
/linux/drivers/dma-buf/ |
A D | udmabuf.c | 25 pgoff_t pagecount; member 67 ret = sg_alloc_table_from_pages(sg, ubuf->pages, ubuf->pagecount, in get_sg_table() 68 0, ubuf->pagecount << PAGE_SHIFT, in get_sg_table() 113 for (pg = 0; pg < ubuf->pagecount; pg++) in release_udmabuf() 189 ubuf->pagecount += list[i].size >> PAGE_SHIFT; in udmabuf_create() 190 if (ubuf->pagecount > pglimit) in udmabuf_create() 193 ubuf->pages = kmalloc_array(ubuf->pagecount, sizeof(*ubuf->pages), in udmabuf_create() 263 exp_info.size = ubuf->pagecount << PAGE_SHIFT; in udmabuf_create()
|
/linux/drivers/hv/ |
A D | channel.c | 314 int pagecount; in create_gpadl_header() local 331 if (pagecount > pfncount) { in create_gpadl_header() 348 pagecount * sizeof(u64); in create_gpadl_header() 357 pfnleft = pagecount - pfncount; in create_gpadl_header() 429 pagecount * sizeof(u64); in create_gpadl_header() 432 for (i = 0; i < pagecount; i++) in create_gpadl_header() 1079 u32 pagecount, void *buffer, u32 bufferlen, in vmbus_sendpacket_pagebuffer() argument 1090 if (pagecount > MAX_PAGE_BUFFER_COUNT) in vmbus_sendpacket_pagebuffer() 1098 ((MAX_PAGE_BUFFER_COUNT - pagecount) * in vmbus_sendpacket_pagebuffer() 1110 desc.rangecount = pagecount; in vmbus_sendpacket_pagebuffer() [all …]
|
/linux/drivers/net/ethernet/google/gve/ |
A D | gve_utils.c | 89 int pagecount = page_count(page_info->page); in gve_dec_pagecnt_bias() local 94 page_info->pagecnt_bias = INT_MAX - pagecount; in gve_dec_pagecnt_bias() 97 page_ref_add(page_info->page, INT_MAX - pagecount); in gve_dec_pagecnt_bias()
|
A D | gve_rx.c | 331 int pagecount = page_count(page_info->page); in gve_rx_can_recycle_buffer() local 334 if (pagecount == page_info->pagecnt_bias) in gve_rx_can_recycle_buffer() 337 else if (pagecount > page_info->pagecnt_bias) in gve_rx_can_recycle_buffer() 339 WARN(pagecount < page_info->pagecnt_bias, in gve_rx_can_recycle_buffer()
|
A D | gve_rx_dqo.c | 383 int pagecount; in gve_try_recycle_buf() local 389 pagecount = gve_buf_ref_cnt(buf_state); in gve_try_recycle_buf() 396 if (pagecount == 1) { in gve_try_recycle_buf()
|
/linux/samples/vfio-mdev/ |
A D | mbochs.c | 155 pgoff_t pagecount; member 175 pgoff_t pagecount; member 534 mdev_state->pages = kcalloc(mdev_state->pagecount, in mbochs_probe() 541 type->name, type->mbytes, mdev_state->pagecount); in mbochs_probe() 724 if (WARN_ON(pgoff >= mdev_state->pagecount)) in mbochs_get_page() 741 for (i = 0; i < mdev_state->pagecount; i++) { in mbochs_put_pages() 757 if (page_offset >= mdev_state->pagecount) in mbochs_region_vm_fault() 795 if (WARN_ON(vmf->pgoff >= dmabuf->pagecount)) in mbochs_dmabuf_vm_fault() 889 for (pg = 0; pg < dmabuf->pagecount; pg++) in mbochs_release_dmabuf() 920 dmabuf->pagecount = DIV_ROUND_UP(mode->size, PAGE_SIZE); in mbochs_dmabuf_alloc() [all …]
|
/linux/drivers/infiniband/hw/hfi1/ |
A D | user_exp_rcv.c | 521 unsigned pagecount, pageidx, setcount = 0, i; in find_phys_blocks() local 535 for (pageidx = 0, pagecount = 1, i = 1; i <= npages; i++) { in find_phys_blocks() 555 while (pagecount) { in find_phys_blocks() 556 int maxpages = pagecount; in find_phys_blocks() 557 u32 bufsize = pagecount * PAGE_SIZE; in find_phys_blocks() 570 pagecount -= maxpages; in find_phys_blocks() 575 pagecount = 1; in find_phys_blocks() 578 pagecount++; in find_phys_blocks()
|
A D | tid_rdma.c | 886 u32 pagecount, pageidx, setcount = 0, i; in tid_rdma_find_phys_blocks_4k() local 899 for (pageidx = 0, pagecount = 1, i = 1; i <= npages; i++) { in tid_rdma_find_phys_blocks_4k() 920 while (pagecount) { in tid_rdma_find_phys_blocks_4k() 921 int maxpages = pagecount; in tid_rdma_find_phys_blocks_4k() 922 u32 bufsize = pagecount * PAGE_SIZE; in tid_rdma_find_phys_blocks_4k() 938 pagecount -= maxpages; in tid_rdma_find_phys_blocks_4k() 943 pagecount = 1; in tid_rdma_find_phys_blocks_4k() 947 pagecount++; in tid_rdma_find_phys_blocks_4k()
|
/linux/arch/x86/hyperv/ |
A D | ivm.c | 258 int hv_set_mem_host_visibility(unsigned long kbuffer, int pagecount, bool visible) in hv_set_mem_host_visibility() argument 273 for (i = 0, pfn = 0; i < pagecount; i++) { in hv_set_mem_host_visibility() 277 if (pfn == HV_MAX_MODIFY_GPA_REP_COUNT || i == pagecount - 1) { in hv_set_mem_host_visibility()
|
/linux/fs/nfs/ |
A D | pagelist.c | 894 unsigned int pagecount, pageused; in nfs_generic_pgio() local 897 pagecount = nfs_page_array_len(mirror->pg_base, mirror->pg_count); in nfs_generic_pgio() 898 pg_array->npages = pagecount; in nfs_generic_pgio() 900 if (pagecount <= ARRAY_SIZE(pg_array->page_array)) in nfs_generic_pgio() 903 pg_array->pagevec = kcalloc(pagecount, sizeof(struct page *), gfp_flags); in nfs_generic_pgio() 922 if (pageused > pagecount) in nfs_generic_pgio() 927 if (WARN_ON_ONCE(pageused != pagecount)) { in nfs_generic_pgio()
|
/linux/drivers/usb/atm/ |
A D | ueagle-atm.c | 724 u8 pagecount, blockcount; in check_dsp_e1() local 729 pagecount = FW_GET_BYTE(dsp); in check_dsp_e1() 733 if (p + 4 * pagecount > len) in check_dsp_e1() 736 for (i = 0; i < pagecount; i++) { in check_dsp_e1() 906 u8 pagecount, blockcount; in uea_load_page_e1() local 921 pagecount = FW_GET_BYTE(p); in uea_load_page_e1() 924 if (pageno >= pagecount) in uea_load_page_e1()
|
/linux/include/linux/ |
A D | hyperv.h | 1173 u32 pagecount,
|