/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_allocator.c | 122 if (array->page_list[p].page) in mthca_array_get() 133 if (!array->page_list[p].page) in mthca_array_set() 136 if (!array->page_list[p].page) in mthca_array_set() 140 ++array->page_list[p].used; in mthca_array_set() 165 array->page_list = kmalloc_array(npage, sizeof(*array->page_list), in mthca_array_init() 167 if (!array->page_list) in mthca_array_init() 185 kfree(array->page_list); in mthca_array_cleanup() 242 if (!buf->page_list) in mthca_buf_alloc() 249 buf->page_list[i].buf = in mthca_buf_alloc() 298 buf->page_list[i].buf, in mthca_buf_free() [all …]
|
A D | mthca_eq.c | 482 eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list), in mthca_create_eq() 484 if (!eq->page_list) in mthca_create_eq() 488 eq->page_list[i].buf = NULL; in mthca_create_eq() 502 if (!eq->page_list[i].buf) in mthca_create_eq() 508 clear_page(eq->page_list[i].buf); in mthca_create_eq() 572 if (eq->page_list[i].buf) in mthca_create_eq() 574 eq->page_list[i].buf, in mthca_create_eq() 575 dma_unmap_addr(&eq->page_list[i], in mthca_create_eq() 581 kfree(eq->page_list); in mthca_create_eq() 621 eq->page_list[i].buf, in mthca_free_eq() [all …]
|
/linux/drivers/xen/ |
A D | unpopulated-alloc.c | 15 static struct page *page_list; variable 89 pg->zone_device_data = page_list; in fill_list() 90 page_list = pg; in fill_list() 124 struct page *pg = page_list; in xen_alloc_unpopulated_pages() 127 page_list = pg->zone_device_data; in xen_alloc_unpopulated_pages() 138 pages[j]->zone_device_data = page_list; in xen_alloc_unpopulated_pages() 139 page_list = pages[j]; in xen_alloc_unpopulated_pages() 165 pages[i]->zone_device_data = page_list; in xen_free_unpopulated_pages() 166 page_list = pages[i]; in xen_free_unpopulated_pages() 195 pg->zone_device_data = page_list; in init() [all …]
|
/linux/mm/ |
A D | dmapool.c | 44 struct list_head page_list; member 55 struct list_head page_list; member 86 list_for_each_entry(page, &pool->page_list, page_list) { in pools_show() 163 INIT_LIST_HEAD(&retval->page_list); in dma_pool_create() 255 list_del(&page->page_list); in pool_free_page() 285 list_for_each_entry_safe(page, tmp, &pool->page_list, page_list) { in dma_pool_destroy() 294 list_del(&page->page_list); in dma_pool_destroy() 325 list_for_each_entry(page, &pool->page_list, page_list) { in dma_pool_alloc() 339 list_add(&page->page_list, &pool->page_list); in dma_pool_alloc() 386 list_for_each_entry(page, &pool->page_list, page_list) { in pool_find_page()
|
A D | vmscan.c | 1529 while (!list_empty(page_list)) { in shrink_page_list() 1538 page = lru_to_page(page_list); in shrink_page_list() 1709 page_list)) in shrink_page_list() 1717 page_list)) in shrink_page_list() 1959 list_splice(&ret_pages, page_list); in shrink_page_list() 1966 struct list_head *page_list) in reclaim_clean_pages_from_list() argument 1998 list_splice(&clean_pages, page_list); in reclaim_clean_pages_from_list() 2393 LIST_HEAD(page_list); in shrink_inactive_list() 2450 free_unref_page_list(&page_list); in shrink_inactive_list() 2610 while (!list_empty(page_list)) { in reclaim_pages() [all …]
|
/linux/drivers/infiniband/hw/usnic/ |
A D | usnic_uiom.c | 53 (void *) &((struct usnic_uiom_chunk *) 0)->page_list[0])) 75 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_put_pages() 89 struct page **page_list; in usnic_uiom_get_pages() local 120 if (!page_list) in usnic_uiom_get_pages() 148 page_list, NULL); in usnic_uiom_get_pages() 157 chunk = kmalloc(struct_size(chunk, page_list, in usnic_uiom_get_pages() 166 sg_init_table(chunk->page_list, chunk->nents); in usnic_uiom_get_pages() 167 for_each_sg(chunk->page_list, sg, chunk->nents, i) { in usnic_uiom_get_pages() 168 sg_set_page(sg, page_list[i + off], in usnic_uiom_get_pages() 191 free_page((unsigned long) page_list); in usnic_uiom_get_pages() [all …]
|
/linux/arch/powerpc/kexec/ |
A D | core_32.c | 32 unsigned long page_list; in default_machine_kexec() local 43 page_list = image->head; in default_machine_kexec() 59 relocate_new_kernel(page_list, reboot_code_buffer_phys, image->start); in default_machine_kexec() 63 (*rnk)(page_list, reboot_code_buffer_phys, image->start); in default_machine_kexec()
|
/linux/drivers/infiniband/core/ |
A D | umem.c | 151 struct page **page_list; in ib_umem_get() local 190 page_list = (struct page **) __get_free_page(GFP_KERNEL); in ib_umem_get() 191 if (!page_list) { in ib_umem_get() 222 gup_flags | FOLL_LONGTERM, page_list); in ib_umem_get() 231 &umem->sgt_append, page_list, pinned, 0, in ib_umem_get() 235 unpin_user_pages_dirty_lock(page_list, pinned, 0); in ib_umem_get() 253 free_page((unsigned long) page_list); in ib_umem_get()
|
/linux/arch/x86/kernel/ |
A D | machine_kexec_32.c | 163 unsigned long page_list[PAGES_NR]; in machine_kexec() local 201 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec() 202 page_list[VA_CONTROL_PAGE] = (unsigned long)control_page; in machine_kexec() 203 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec() 206 page_list[PA_SWAP_PAGE] = (page_to_pfn(image->swap_page) in machine_kexec() 229 (unsigned long)page_list, in machine_kexec()
|
A D | machine_kexec_64.c | 299 unsigned long page_list[PAGES_NR]; in machine_kexec() local 330 page_list[PA_CONTROL_PAGE] = virt_to_phys(control_page); in machine_kexec() 331 page_list[VA_CONTROL_PAGE] = (unsigned long)control_page; in machine_kexec() 332 page_list[PA_TABLE_PAGE] = in machine_kexec() 336 page_list[PA_SWAP_PAGE] = (page_to_pfn(image->swap_page) in machine_kexec() 359 (unsigned long)page_list, in machine_kexec()
|
/linux/drivers/comedi/ |
A D | comedi_buf.c | 29 if (bm->page_list) { in comedi_buf_map_kref_release() 35 buf = &bm->page_list[0]; in comedi_buf_map_kref_release() 41 buf = &bm->page_list[i]; in comedi_buf_map_kref_release() 46 vfree(bm->page_list); in comedi_buf_map_kref_release() 93 bm->page_list = vzalloc(sizeof(*buf) * n_pages); in comedi_buf_map_alloc() 94 if (!bm->page_list) in comedi_buf_map_alloc() 112 buf = &bm->page_list[i]; in comedi_buf_map_alloc() 120 buf = &bm->page_list[i]; in comedi_buf_map_alloc() 170 buf = &bm->page_list[0]; in __comedi_buf_alloc() 178 buf = &bm->page_list[i]; in __comedi_buf_alloc() [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | mr.c | 693 int start_index, int npages, u64 *page_list) in mlx4_write_mtt_chunk() argument 719 int start_index, int npages, u64 *page_list) in __mlx4_write_mtt() argument 739 page_list += chunk; in __mlx4_write_mtt() 747 int start_index, int npages, u64 *page_list) in mlx4_write_mtt() argument 770 inbox[i + 2] = cpu_to_be64(page_list[i] | in mlx4_write_mtt() 780 page_list += chunk; in mlx4_write_mtt() 793 u64 *page_list; in mlx4_buf_write_mtt() local 797 page_list = kcalloc(buf->npages, sizeof(*page_list), GFP_KERNEL); in mlx4_buf_write_mtt() 798 if (!page_list) in mlx4_buf_write_mtt() 805 page_list[i] = buf->page_list[i].map; in mlx4_buf_write_mtt() [all …]
|
A D | alloc.c | 619 buf->page_list = kcalloc(buf->nbufs, sizeof(*buf->page_list), in mlx4_buf_alloc() 621 if (!buf->page_list) in mlx4_buf_alloc() 625 buf->page_list[i].buf = in mlx4_buf_alloc() 628 if (!buf->page_list[i].buf) in mlx4_buf_alloc() 631 buf->page_list[i].map = t; in mlx4_buf_alloc() 653 if (buf->page_list[i].buf) in mlx4_buf_free() 656 buf->page_list[i].buf, in mlx4_buf_free() 657 buf->page_list[i].map); in mlx4_buf_free() 658 kfree(buf->page_list); in mlx4_buf_free()
|
A D | eq.c | 989 eq->page_list = kmalloc_array(npages, sizeof(*eq->page_list), in mlx4_create_eq() 991 if (!eq->page_list) in mlx4_create_eq() 1011 if (!eq->page_list[i].buf) in mlx4_create_eq() 1015 eq->page_list[i].map = t; in mlx4_create_eq() 1072 if (eq->page_list[i].buf) in mlx4_create_eq() 1074 eq->page_list[i].buf, in mlx4_create_eq() 1075 eq->page_list[i].map); in mlx4_create_eq() 1080 kfree(eq->page_list); in mlx4_create_eq() 1108 eq->page_list[i].buf, in mlx4_free_eq() 1109 eq->page_list[i].map); in mlx4_free_eq() [all …]
|
/linux/drivers/misc/genwqe/ |
A D | card_utils.c | 243 struct page **page_list, int num_pages, in genwqe_map_pages() argument 254 daddr = dma_map_page(&pci_dev->dev, page_list[i], in genwqe_map_pages() 568 m->page_list = kcalloc(m->nr_pages, in genwqe_user_vmap() 571 if (!m->page_list) { in genwqe_user_vmap() 584 m->page_list); /* ptrs to pages */ in genwqe_user_vmap() 605 kfree(m->page_list); in genwqe_user_vmap() 606 m->page_list = NULL; in genwqe_user_vmap() 633 if (m->page_list) { in genwqe_user_vunmap() 634 unpin_user_pages_dirty_lock(m->page_list, m->nr_pages, in genwqe_user_vunmap() 636 kfree(m->page_list); in genwqe_user_vunmap() [all …]
|
/linux/include/linux/ |
A D | dm-io.h | 23 struct page_list { struct 24 struct page_list *next; argument 43 struct page_list *pl; argument
|
/linux/drivers/md/ |
A D | dm-kcopyd.c | 58 struct page_list *pages; 93 static struct page_list zero_page_list; 214 static struct page_list *alloc_pl(gfp_t gfp) in alloc_pl() 216 struct page_list *pl; in alloc_pl() 231 static void free_pl(struct page_list *pl) in free_pl() 243 struct page_list *next; in kcopyd_put_pages() 263 struct page_list *pl; in kcopyd_get_pages() 292 static void drop_pages(struct page_list *pl) in drop_pages() 294 struct page_list *next; in drop_pages() 309 struct page_list *pl = NULL, *next; in client_reserve_pages() [all …]
|
A D | dm-io.c | 183 struct page_list *pl = (struct page_list *) dp->context_ptr; in list_get_page() 192 struct page_list *pl = (struct page_list *) dp->context_ptr; in list_next_page() 197 static void list_dp_init(struct dpages *dp, struct page_list *pl, unsigned offset) in list_dp_init()
|
/linux/arch/arm/kernel/ |
A D | machine_kexec.c | 152 unsigned long page_list, reboot_entry_phys; in machine_kexec() local 165 page_list = image->head & PAGE_MASK; in machine_kexec() 176 data->kexec_indirection_page = page_list; in machine_kexec()
|
/linux/arch/sh/kernel/ |
A D | machine_kexec.c | 71 unsigned long page_list; in machine_kexec() local 101 page_list = image->head; in machine_kexec() 118 (*rnk)(page_list, reboot_code_buffer, in machine_kexec()
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_validation.h | 80 struct list_head page_list; member 112 .page_list = LIST_HEAD_INIT((_name).page_list), \
|
/linux/Documentation/admin-guide/device-mapper/ |
A D | dm-io.rst | 25 struct page_list { 26 struct page_list *next; 31 struct page_list *pl, unsigned int offset, 34 struct page_list *pl, unsigned int offset,
|
/linux/mm/damon/ |
A D | paddr.c | 220 LIST_HEAD(page_list); in damon_pa_apply_scheme() 240 list_add(&page->lru, &page_list); in damon_pa_apply_scheme() 244 reclaim_pages(&page_list); in damon_pa_apply_scheme()
|
/linux/include/uapi/linux/ |
A D | vbox_vmmdev_types.h | 238 } page_list; member 261 } page_list; member
|
/linux/drivers/vhost/ |
A D | vdpa.c | 689 struct page **page_list; in vhost_vdpa_pa_map() local 699 page_list = (struct page **) __get_free_page(GFP_KERNEL); in vhost_vdpa_pa_map() 700 if (!page_list) in vhost_vdpa_pa_map() 727 gup_flags, page_list, NULL); in vhost_vdpa_pa_map() 732 unpin_user_pages(page_list, pinned); in vhost_vdpa_pa_map() 740 map_pfn = page_to_pfn(page_list[0]); in vhost_vdpa_pa_map() 743 unsigned long this_pfn = page_to_pfn(page_list[i]); in vhost_vdpa_pa_map() 761 unpin_user_pages(&page_list[i], in vhost_vdpa_pa_map() 804 free_page((unsigned long)page_list); in vhost_vdpa_pa_map()
|