Home
last modified time | relevance | path

Searched refs:__pa (Results 1 – 25 of 421) sorted by relevance

12345678910>>...17

/linux/arch/xtensa/kernel/
A Dsetup.c298 !mem_reserve(__pa(initrd_start), __pa(initrd_end))) in setup_arch()
304 mem_reserve(__pa(_stext), __pa(_end)); in setup_arch()
306 mem_reserve(__pa(_xip_start), __pa(_xip_end)); in setup_arch()
311 mem_reserve(__pa(_WindowVectors_text_start), in setup_arch()
312 __pa(_WindowVectors_text_end)); in setup_arch()
316 __pa(_DebugInterruptVector_text_end)); in setup_arch()
322 __pa(_UserExceptionVector_text_end)); in setup_arch()
327 mem_reserve(__pa(_exception_text_start), in setup_arch()
328 __pa(_exception_text_end)); in setup_arch()
354 __pa(_SecondaryResetVector_text_end)); in setup_arch()
[all …]
/linux/arch/parisc/kernel/
A Dfirmware.c162 __pa(pdc_result), 0); in set_firmware_width_unlocked()
268 __pa(pdc_result), __pa(pdc_result2), len); in pdc_chassis_info()
597 __pa(orig_prod_num), __pa(current_prod_num), __pa(serial_no)); in pdc_model_platform_info()
689 __pa(pdc_result2)); in pdc_mem_map_hpa()
863 __pa(pdc_result), __pa(hwpath)); in pdc_get_initiator()
946 __pa(pdc_result), hpa, __pa(tbl)); in pdc_pci_irt()
1287 __pa(iodc_retbuf), 0, __pa(iodc_dbuf), i, 0); in pdc_iodc_print()
1315 __pa(iodc_retbuf), 0, __pa(iodc_dbuf), 1, 0); in pdc_iodc_getc()
1412 __pa(pdc_result), __pa(&result), *actcnt, in pdc_pat_cell_info()
1525 __pa(pdc_result)); in pdc_pat_pd_get_pdc_revisions()
[all …]
/linux/arch/x86/include/asm/
A Dpgalloc.h67 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel()
68 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in pmd_populate_kernel()
74 paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); in pmd_populate_kernel_safe()
101 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate()
102 set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); in pud_populate()
107 paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); in pud_populate_safe()
115 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate()
116 set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); in p4d_populate()
121 paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); in p4d_populate_safe()
138 paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); in pgd_populate()
[all …]
A Dpage.h41 #ifndef __pa
42 #define __pa(x) __phys_addr((unsigned long)(x)) macro
63 #define __boot_pa(x) __pa(x)
69 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT)
/linux/arch/x86/kernel/
A Dhead32.c74 #ifdef __pa in mk_early_pgtbl_32()
75 #undef __pa in mk_early_pgtbl_32()
77 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET) in mk_early_pgtbl_32() macro
82 const unsigned long limit = __pa(_end) + in mk_early_pgtbl_32()
85 pmd_t pl2, *pl2p = (pmd_t *)__pa(initial_pg_pmd); in mk_early_pgtbl_32()
88 pgd_t pl2, *pl2p = (pgd_t *)__pa(initial_page_table); in mk_early_pgtbl_32()
92 ptep = (pte_t *)__pa(__brk_base); in mk_early_pgtbl_32()
112 ptr = (unsigned long *)__pa(&max_pfn_mapped); in mk_early_pgtbl_32()
116 ptr = (unsigned long *)__pa(&_brk_end); in mk_early_pgtbl_32()
A Dmachine_kexec_32.c89 set_pgd(pgd, __pgd(__pa(pmd) | _PAGE_PRESENT)); in machine_kexec_page_table_set_one()
95 set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); in machine_kexec_page_table_set_one()
111 (unsigned long)control_page, __pa(control_page)); in machine_kexec_prepare_page_tables()
117 __pa(control_page), __pa(control_page)); in machine_kexec_prepare_page_tables()
201 page_list[PA_CONTROL_PAGE] = __pa(control_page); in machine_kexec()
203 page_list[PA_PGD] = __pa(image->arch.pgd); in machine_kexec()
A Despfix_64.c170 pud = __pud(__pa(pmd_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
171 paravirt_alloc_pmd(&init_mm, __pa(pmd_p) >> PAGE_SHIFT); in init_espfix_ap()
182 pmd = __pmd(__pa(pte_p) | (PGTABLE_PROT & ptemask)); in init_espfix_ap()
183 paravirt_alloc_pte(&init_mm, __pa(pte_p) >> PAGE_SHIFT); in init_espfix_ap()
194 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap()
/linux/arch/arm/mach-omap2/
A Domap-secure.c76 outer_clean_range(__pa(param), __pa(param + 5)); in omap_secure_dispatcher()
77 ret = omap_smc2(idx, flag, __pa(param)); in omap_secure_dispatcher()
129 param[1] = __pa(addr); /* Physical address for saving */ in omap3_save_secure_ram()
134 ret = save_secure_ram_context(__pa(param)); in omap3_save_secure_ram()
172 outer_clean_range(__pa(param), __pa(param + 5)); in rx51_secure_dispatcher()
173 ret = omap_smc3(idx, process, flag, __pa(param)); in rx51_secure_dispatcher()
/linux/arch/csky/mm/
A Dinit.c55 if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) { in setup_initrd()
62 if (memblock_is_region_reserved(__pa(initrd_start), size)) { in setup_initrd()
64 __pa(initrd_start), size); in setup_initrd()
68 memblock_reserve(__pa(initrd_start), size); in setup_initrd()
122 p[i] = __pa(invalid_pte_table); in pgd_init()
133 swapper_pg_dir[i].pgd = __pa(invalid_pte_table); in mmu_init()
137 __pa(kernel_pte_tables + (PTRS_PER_PTE * (i - USER_PTRS_PER_PGD))); in mmu_init()
182 set_pmd(pmd, __pmd(__pa(pte))); in fixrange_init()
/linux/arch/parisc/mm/
A Dinit.c284 memblock_reserve(__pa(KERNEL_BINARY_TEXT_START), in setup_bootmem()
300 if (__pa(initrd_start) < mem_max) { in setup_bootmem()
303 if (__pa(initrd_end) > mem_max) { in setup_bootmem()
304 initrd_reserve = mem_max - __pa(initrd_start); in setup_bootmem()
309 …printk(KERN_INFO "initrd: reserving %08lx-%08lx (mem_max %08lx)\n", __pa(initrd_start), __pa(initr… in setup_bootmem()
357 ro_start = __pa((unsigned long)_text); in map_pages()
358 ro_end = __pa((unsigned long)&data_start); in map_pages()
359 kernel_start = __pa((unsigned long)&__init_begin); in map_pages()
360 kernel_end = __pa((unsigned long)&_end); in map_pages()
444 map_pages(start, __pa(start), end-start, in set_kernel_text_rw()
[all …]
/linux/arch/um/kernel/
A Dphysmem.c107 os_seek_file(physmem_fd, __pa(__syscall_stub_start)); in setup_physmem()
111 memblock_add(__pa(start), len + highmem); in setup_physmem()
112 memblock_reserve(__pa(start), reserve); in setup_physmem()
114 min_low_pfn = PFN_UP(__pa(reserve_end)); in setup_physmem()
126 else if (phys < __pa(end_iomem)) { in phys_mapping()
139 else if (phys < __pa(end_iomem) + highmem) { in phys_mapping()
212 region->phys = __pa(region->virt); in setup_iomem()
/linux/arch/x86/xen/
A Dp2m.c386 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL)); in xen_rebuild_p2m_list()
396 pfn_pte(PFN_DOWN(__pa(mfns)), PAGE_KERNEL_RO)); in xen_rebuild_p2m_list()
406 set_pmd(pmdp, __pmd(__pa(ptep) | _KERNPG_TABLE)); in xen_rebuild_p2m_list()
455 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_identity))) in get_phys_to_machine()
505 __pmd(__pa(pte_newpg[i]) | _KERNPG_TABLE)); in alloc_p2m_pmd()
586 if (p2m_pfn == PFN_DOWN(__pa(p2m_identity)) || in xen_alloc_p2m_entry()
587 p2m_pfn == PFN_DOWN(__pa(p2m_missing))) { in xen_alloc_p2m_entry()
595 if (p2m_pfn == PFN_DOWN(__pa(p2m_missing))) in xen_alloc_p2m_entry()
606 pfn_pte(PFN_DOWN(__pa(p2m)), PAGE_KERNEL)); in xen_alloc_p2m_entry()
669 if (pte_pfn(*ptep) == PFN_DOWN(__pa(p2m_missing))) in __set_phys_to_machine()
[all …]
/linux/arch/powerpc/mm/nohash/
A D8xx.c38 return __pa(va); in v_block_mapped()
150 unsigned long etext8 = ALIGN(__pa(_etext), SZ_8M); in mmu_mapin_ram()
151 unsigned long sinittext = __pa(_sinittext); in mmu_mapin_ram()
154 unsigned long einittext8 = ALIGN(__pa(_einittext), SZ_8M); in mmu_mapin_ram()
181 unsigned long etext8 = ALIGN(__pa(_etext), SZ_8M); in mmu_mark_initmem_nx()
182 unsigned long sinittext = __pa(_sinittext); in mmu_mark_initmem_nx()
184 unsigned long einittext8 = ALIGN(__pa(_einittext), SZ_8M); in mmu_mark_initmem_nx()
195 unsigned long sinittext = __pa(_sinittext); in mmu_mark_rodata_ro()
/linux/arch/arm/mm/
A Dpmsa-v8.c255 subtract_range(mem, ARRAY_SIZE(mem), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
256 subtract_range(io, ARRAY_SIZE(io), __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
260 subtract_range(mem, ARRAY_SIZE(mem), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
261 subtract_range(io, ARRAY_SIZE(io), CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
277 err |= pmsav8_setup_fixed(PMSAv8_XIP_REGION, CONFIG_XIP_PHYS_ADDR, __pa(_exiprom)); in pmsav8_setup()
280 err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END)); in pmsav8_setup()
/linux/arch/riscv/mm/
A Dinit.c203 max_mapped_addr = __pa(~(ulong)0); in setup_bootmem()
309 return __pa(vaddr); in alloc_pte_late()
371 return __pa(vaddr); in alloc_pmd_late()
403 #define get_pgd_next_virt(__pa) pt_ops.get_pmd_virt(__pa) argument
405 create_pmd_mapping(__nextp, __va, __pa, __sz, __prot)
410 #define get_pgd_next_virt(__pa) pt_ops.get_pte_virt(__pa) argument
723 if (start <= __pa(PAGE_OFFSET) && in setup_vm_final()
724 __pa(PAGE_OFFSET) < end) in setup_vm_final()
725 start = __pa(PAGE_OFFSET); in setup_vm_final()
726 if (end >= __pa(PAGE_OFFSET) + memory_limit) in setup_vm_final()
[all …]
A Dkasan_init.c31 (__pa((uintptr_t) kasan_early_shadow_pte)), in kasan_early_init()
38 (__pa(((uintptr_t) kasan_early_shadow_pmd))), in kasan_early_init()
48 (__pa(((uintptr_t) kasan_early_shadow_pmd))), in kasan_early_init()
73 set_pmd(pmd, pfn_pmd(PFN_DOWN(__pa(base_pte)), PAGE_TABLE)); in kasan_populate_pte()
108 set_pgd(pgd, pfn_pgd(PFN_DOWN(__pa(base_pmd)), PAGE_TABLE)); in kasan_populate_pmd()
159 set_pgd(pgd_k, pfn_pgd(PFN_DOWN(__pa(p)), PAGE_TABLE)); in kasan_shallow_populate_pgd()
/linux/arch/x86/power/
A Dhibernate_64.c71 __pud(__pa(pmd) | pgprot_val(pgtable_prot))); in set_up_temporary_text_mapping()
73 p4d_t new_p4d = __p4d(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping()
74 pgd_t new_pgd = __pgd(__pa(p4d) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping()
80 pgd_t new_pgd = __pgd(__pa(pud) | pgprot_val(pgtable_prot)); in set_up_temporary_text_mapping()
123 temp_pgt = __pa(pgd); in set_up_temporary_mappings()
A Dhibernate_32.c41 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in resume_one_md_table_init()
66 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in resume_one_page_table_init()
141 __pgd(__pa(empty_zero_page) | _PAGE_PRESENT)); in resume_init_first_level_page_table()
189 temp_pgt = __pa(resume_pg_dir); in swsusp_arch_resume()
/linux/arch/s390/mm/
A Dkasan_init.c178 pmd_val(*pm_dir) = __pa(page) | sgt_prot; in kasan_early_pgtable_populate()
197 pte_val(*pt_dir) = __pa(page) | pgt_prot; in kasan_early_pgtable_populate()
202 pte_val(*pt_dir) = __pa(page) | pgt_prot; in kasan_early_pgtable_populate()
206 pte_val(*pt_dir) = __pa(page) | pgt_prot_zero; in kasan_early_pgtable_populate()
222 S390_lowcore.kernel_asce = (__pa(pgd) & PAGE_MASK) | asce_bits; in kasan_set_pgd()
259 pmd_t pmd_z = __pmd(__pa(kasan_early_shadow_pte) | _SEGMENT_ENTRY); in kasan_early_init()
260 pud_t pud_z = __pud(__pa(kasan_early_shadow_pmd) | _REGION3_ENTRY); in kasan_early_init()
261 p4d_t p4d_z = __p4d(__pa(kasan_early_shadow_pud) | _REGION2_ENTRY); in kasan_early_init()
266 pte_z = __pte(__pa(kasan_early_shadow_page) | pgt_prot); in kasan_early_init()
/linux/arch/hexagon/include/asm/
A Dpage.h86 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + PHYS_OFFSET) macro
93 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr)))
99 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT)
129 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
/linux/arch/x86/mm/
A Dinit_32.c77 paravirt_alloc_pmd(&init_mm, __pa(pmd_table) >> PAGE_SHIFT); in one_md_table_init()
78 set_pgd(pgd, __pgd(__pa(pmd_table) | _PAGE_PRESENT)); in one_md_table_init()
102 paravirt_alloc_pte(&init_mm, __pa(page_table) >> PAGE_SHIFT); in one_page_table_init()
103 set_pmd(pmd, __pmd(__pa(page_table) | _PAGE_TABLE)); in one_page_table_init()
184 paravirt_alloc_pte(&init_mm, __pa(newpte) >> PAGE_SHIFT); in page_table_kmap_check()
185 set_pmd(pmd, __pmd(__pa(newpte)|_PAGE_TABLE)); in page_table_kmap_check()
189 paravirt_release_pte(__pa(pte) >> PAGE_SHIFT); in page_table_kmap_check()
473 pfn, pmd, __pa(pmd)); in native_pagetable_init()
482 pfn, pmd, __pa(pmd), pte, __pa(pte)); in native_pagetable_init()
485 paravirt_alloc_pmd(&init_mm, __pa(base) >> PAGE_SHIFT); in native_pagetable_init()
A Dkasan_init_64.c29 __pa(MAX_DMA_ADDRESS), MEMBLOCK_ALLOC_ACCESSIBLE, nid); in early_alloc()
33 (void *)_RET_IP_, nid, __pa(MAX_DMA_ADDRESS)); in early_alloc()
50 if (p && pmd_set_huge(pmd, __pa(p), PAGE_KERNEL)) in kasan_populate_pmd()
68 entry = pfn_pte(PFN_DOWN(__pa(p)), PAGE_KERNEL); in kasan_populate_pmd()
86 if (p && pud_set_huge(pud, __pa(p), PAGE_KERNEL)) in kasan_populate_pud()
345 __pgd(__pa(tmp_p4d_table) | _KERNPG_TABLE)); in kasan_init()
404 early_pfn_to_nid(__pa(_stext))); in kasan_init()
425 pte = __pte(__pa(kasan_early_shadow_page) | pgprot_val(prot)); in kasan_init()
/linux/drivers/i2c/busses/
A Di2c-opal.c90 req.buffer_ra = cpu_to_be64(__pa(msgs[0].buf)); in i2c_opal_master_xfer()
101 req.buffer_ra = cpu_to_be64(__pa(msgs[1].buf)); in i2c_opal_master_xfer()
126 req.buffer_ra = cpu_to_be64(__pa(&data->byte)); in i2c_opal_smbus_xfer()
134 req.buffer_ra = cpu_to_be64(__pa(&data->byte)); in i2c_opal_smbus_xfer()
146 req.buffer_ra = cpu_to_be64(__pa(local)); in i2c_opal_smbus_xfer()
154 req.buffer_ra = cpu_to_be64(__pa(&data->block[1])); in i2c_opal_smbus_xfer()
/linux/arch/ia64/include/asm/
A Dpgalloc.h36 p4d_val(*p4d_entry) = __pa(pud); in p4d_populate()
45 pud_val(*pud_entry) = __pa(pmd); in pud_populate()
59 pmd_val(*pmd_entry) = __pa(pte); in pmd_populate_kernel()
/linux/arch/powerpc/include/asm/
A Dpage.h131 #define virt_to_pfn(kaddr) (__pa(kaddr) >> PAGE_SHIFT)
212 #define __pa(x) ((phys_addr_t)(unsigned long)(x) - VIRT_PHYS_OFFSET) macro
226 #define __pa(x) \ macro
234 #define __pa(x) ((unsigned long)(x) - PAGE_OFFSET + MEMORY_START) macro

Completed in 39 milliseconds

12345678910>>...17