/linux/include/trace/events/ |
A D | fs_dax.h | 65 struct page *zero_page, 67 TP_ARGS(inode, vmf, zero_page, radix_entry), 72 __field(struct page *, zero_page) 81 __entry->zero_page = zero_page; 91 __entry->zero_page, 99 struct page *zero_page, void *radix_entry), \ 100 TP_ARGS(inode, vmf, zero_page, radix_entry))
|
/linux/arch/nds32/mm/ |
A D | init.c | 140 void *zero_page; in paging_init() local 152 zero_page = memblock_alloc(PAGE_SIZE, PAGE_SIZE); in paging_init() 153 if (!zero_page) in paging_init() 158 empty_zero_page = virt_to_page(zero_page); in paging_init()
|
/linux/arch/arm64/kernel/ |
A D | hibernate.c | 412 void *zero_page; in swsusp_arch_resume() local 437 zero_page = (void *)get_safe_page(GFP_ATOMIC); in swsusp_arch_resume() 438 if (!zero_page) { in swsusp_arch_resume() 474 resume_hdr.__hyp_stub_vectors, virt_to_phys(zero_page)); in swsusp_arch_resume()
|
A D | machine_kexec.c | 150 kimage->arch.zero_page = __pa_symbol(empty_zero_page); in machine_kexec_post_load()
|
A D | asm-offsets.c | 178 DEFINE(KIMAGE_ARCH_ZERO_PAGE, offsetof(struct kimage, arch.zero_page)); in main()
|
/linux/mm/ |
A D | huge_memory.c | 93 struct page *zero_page; in get_huge_zero_page() local 100 if (!zero_page) { in get_huge_zero_page() 106 if (cmpxchg(&huge_zero_page, NULL, zero_page)) { in get_huge_zero_page() 108 __free_pages(zero_page, compound_order(zero_page)); in get_huge_zero_page() 160 BUG_ON(zero_page == NULL); in shrink_huge_zero_page_scan() 162 __free_pages(zero_page, compound_order(zero_page)); in shrink_huge_zero_page_scan() 711 struct page *zero_page) in set_huge_zero_page() argument 716 entry = mk_pmd(zero_page, vma->vm_page_prot); in set_huge_zero_page() 741 struct page *zero_page; in do_huge_pmd_anonymous_page() local 747 if (unlikely(!zero_page)) { in do_huge_pmd_anonymous_page() [all …]
|
/linux/arch/arm64/include/asm/ |
A D | kexec.h | 108 phys_addr_t zero_page; member
|
A D | assembler.h | 468 .macro break_before_make_ttbr_switch zero_page, page_table, tmp, tmp2 469 phys_to_ttbr \tmp, \zero_page
|
/linux/drivers/dma/ |
A D | bcm2835-dma.c | 50 dma_addr_t zero_page; member 750 if (buf_addr == od->zero_page && !c->is_lite_channel) in bcm2835_dma_prep_dma_cyclic() 852 dma_unmap_page_attrs(od->ddev.dev, od->zero_page, PAGE_SIZE, in bcm2835_dma_free() 937 od->zero_page = dma_map_page_attrs(od->ddev.dev, ZERO_PAGE(0), 0, in bcm2835_dma_probe() 940 if (dma_mapping_error(od->ddev.dev, od->zero_page)) { in bcm2835_dma_probe()
|
/linux/fs/ |
A D | dax.c | 1078 struct page *zero_page; in dax_pmd_load_hole() local 1083 zero_page = mm_get_huge_zero_page(vmf->vma->vm_mm); in dax_pmd_load_hole() 1085 if (unlikely(!zero_page)) in dax_pmd_load_hole() 1088 pfn = page_to_pfn_t(zero_page); in dax_pmd_load_hole() 1108 pmd_entry = mk_pmd(zero_page, vmf->vma->vm_page_prot); in dax_pmd_load_hole() 1112 trace_dax_pmd_load_hole(inode, vmf, zero_page, *entry); in dax_pmd_load_hole() 1118 trace_dax_pmd_load_hole_fallback(inode, vmf, zero_page, *entry); in dax_pmd_load_hole()
|
/linux/include/target/ |
A D | target_core_fabric.h | 215 u32 length, bool zero_page, bool chainable);
|
/linux/drivers/nvdimm/ |
A D | pfn_devs.c | 364 void *zero_page = page_address(ZERO_PAGE(0)); in nd_pfn_clear_memmap_errors() local 402 rc = nvdimm_write_bytes(ndns, nsoff, zero_page, in nd_pfn_clear_memmap_errors()
|
A D | btt.c | 511 void *zero_page = page_address(ZERO_PAGE(0)); in arena_clear_freelist_error() local 521 ret = arena_write_bytes(arena, nsoff, zero_page, in arena_clear_freelist_error()
|
/linux/arch/arm/mm/ |
A D | mmu.c | 1724 void *zero_page; in paging_init() local 1747 zero_page = early_alloc(PAGE_SIZE); in paging_init() 1751 empty_zero_page = virt_to_page(zero_page); in paging_init()
|
/linux/drivers/target/ |
A D | target_core_transport.c | 2729 bool zero_page, bool chainable) in target_alloc_sgl() argument 2731 gfp_t gfp = GFP_KERNEL | (zero_page ? __GFP_ZERO : 0); in target_alloc_sgl()
|
/linux/virt/kvm/ |
A D | kvm_main.c | 3003 const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0))); in kvm_clear_guest() local 3010 ret = kvm_write_guest_page(kvm, gfn, zero_page, offset, len); in kvm_clear_guest()
|
/linux/arch/x86/kvm/vmx/ |
A D | vmx.c | 3556 const void *zero_page = (const void *) __va(page_to_phys(ZERO_PAGE(0))); in init_rmode_tss() local 3561 if (__copy_to_user(ua + PAGE_SIZE * i, zero_page, PAGE_SIZE)) in init_rmode_tss()
|