/linux/kernel/dma/ |
A D | ops_helpers.c | 8 static struct page *dma_common_vaddr_to_page(void *cpu_addr) in dma_common_vaddr_to_page() argument 10 if (is_vmalloc_addr(cpu_addr)) in dma_common_vaddr_to_page() 11 return vmalloc_to_page(cpu_addr); in dma_common_vaddr_to_page() 12 return virt_to_page(cpu_addr); in dma_common_vaddr_to_page() 19 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_common_get_sgtable() argument 22 struct page *page = dma_common_vaddr_to_page(cpu_addr); in dma_common_get_sgtable() 35 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_common_mmap() argument 42 struct page *page = dma_common_vaddr_to_page(cpu_addr); in dma_common_mmap() 47 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in dma_common_mmap()
|
A D | remap.c | 9 struct page **dma_common_find_pages(void *cpu_addr) in dma_common_find_pages() argument 11 struct vm_struct *area = find_vm_area(cpu_addr); in dma_common_find_pages() 60 void dma_common_free_remap(void *cpu_addr, size_t size) in dma_common_free_remap() argument 62 struct vm_struct *area = find_vm_area(cpu_addr); in dma_common_free_remap() 65 WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); in dma_common_free_remap() 69 vunmap(cpu_addr); in dma_common_free_remap()
|
A D | mapping.c | 389 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_get_sgtable_attrs() argument 453 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_mmap_attrs() argument 459 return dma_direct_mmap(dev, vma, cpu_addr, dma_addr, size, in dma_mmap_attrs() 492 void *cpu_addr; in dma_alloc_attrs() local 497 return cpu_addr; in dma_alloc_attrs() 505 cpu_addr = ops->alloc(dev, size, dma_handle, flag, attrs); in dma_alloc_attrs() 510 return cpu_addr; in dma_alloc_attrs() 530 if (!cpu_addr) in dma_free_attrs() 533 debug_dma_free_coherent(dev, size, cpu_addr, dma_handle); in dma_free_attrs() 535 dma_direct_free(dev, size, cpu_addr, dma_handle, attrs); in dma_free_attrs() [all …]
|
A D | direct.c | 275 void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs) in dma_direct_free() argument 282 dma_free_contiguous(dev, cpu_addr, size); in dma_direct_free() 291 arch_dma_free(dev, size, cpu_addr, dma_addr, attrs); in dma_direct_free() 297 if (!dma_release_from_global_coherent(page_order, cpu_addr)) in dma_direct_free() 304 dma_free_from_pool(dev, cpu_addr, PAGE_ALIGN(size))) in dma_direct_free() 310 if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) in dma_direct_free() 311 vunmap(cpu_addr); in dma_direct_free() 313 arch_dma_clear_uncached(cpu_addr, size); in dma_direct_free() 474 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_direct_get_sgtable() argument 493 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_direct_mmap() argument [all …]
|
/linux/drivers/pci/controller/dwc/ |
A D | pcie-designware.c | 278 lower_32_bits(cpu_addr)); in dw_pcie_prog_outbound_atu_unroll() 280 upper_32_bits(cpu_addr)); in dw_pcie_prog_outbound_atu_unroll() 320 cpu_addr = pci->ops->cpu_addr_fixup(pci, cpu_addr); in __dw_pcie_prog_outbound_atu() 324 cpu_addr, pci_addr, size); in __dw_pcie_prog_outbound_atu() 331 lower_32_bits(cpu_addr)); in __dw_pcie_prog_outbound_atu() 333 upper_32_bits(cpu_addr)); in __dw_pcie_prog_outbound_atu() 369 cpu_addr, pci_addr, size); in dw_pcie_prog_outbound_atu() 403 lower_32_bits(cpu_addr)); in dw_pcie_prog_inbound_atu_unroll() 405 upper_32_bits(cpu_addr)); in dw_pcie_prog_inbound_atu_unroll() 443 int bar, u64 cpu_addr, in dw_pcie_prog_inbound_atu() argument [all …]
|
A D | pcie-tegra194-acpi.c | 48 int index, int type, u64 cpu_addr, in program_outbound_atu() argument 51 atu_reg_write(pcie_ecam, index, lower_32_bits(cpu_addr), in program_outbound_atu() 53 atu_reg_write(pcie_ecam, index, upper_32_bits(cpu_addr), in program_outbound_atu() 57 atu_reg_write(pcie_ecam, index, lower_32_bits(cpu_addr + size - 1), in program_outbound_atu()
|
/linux/drivers/pci/controller/cadence/ |
A D | pcie-cadence-host.c | 185 u64 cpu_addr, u64 size, in cdns_pcie_host_bar_ib_config() argument 199 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_host_bar_ib_config() 212 if (size + cpu_addr >= SZ_4G) { in cdns_pcie_host_bar_ib_config() 279 u64 cpu_addr, pci_addr, size, winsize; in cdns_pcie_host_bar_config() local 286 cpu_addr = entry->res->start; in cdns_pcie_host_bar_config() 293 pci_addr, cpu_addr); in cdns_pcie_host_bar_config() 329 cpu_addr); in cdns_pcie_host_bar_config() 342 cpu_addr += winsize; in cdns_pcie_host_bar_config() 402 u64 cpu_addr = cfg_res->start; in cdns_pcie_host_init_address_translation() local 421 cpu_addr = pcie->ops->cpu_addr_fixup(pcie, cpu_addr); in cdns_pcie_host_init_address_translation() [all …]
|
A D | pcie-cadence.c | 28 u64 cpu_addr, u64 pci_addr, size_t size) in cdns_pcie_set_outbound_region() argument 93 cpu_addr = pcie->ops->cpu_addr_fixup(pcie, cpu_addr); in cdns_pcie_set_outbound_region() 96 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_set_outbound_region() 97 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_set_outbound_region() 105 u32 r, u64 cpu_addr) in cdns_pcie_set_outbound_region_for_normal_msg() argument 123 cpu_addr = pcie->ops->cpu_addr_fixup(pcie, cpu_addr); in cdns_pcie_set_outbound_region_for_normal_msg() 126 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_set_outbound_region_for_normal_msg() 127 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_set_outbound_region_for_normal_msg()
|
/linux/arch/s390/kvm/ |
A D | trace.h | 220 TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr, \ 222 TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr, parameter), 227 __field(__u16, cpu_addr) 234 __entry->cpu_addr = cpu_addr; 242 __entry->cpu_addr, __entry->parameter) 246 TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr), 247 TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr), 252 __field(__u16, cpu_addr) 258 __entry->cpu_addr = cpu_addr; 265 __entry->cpu_addr)
|
A D | sigp.c | 271 u16 cpu_addr, u32 parameter, u64 *status_reg) in handle_sigp_dst() argument 274 struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr); in handle_sigp_dst() 348 u16 cpu_addr) in handle_sigp_order_in_user_space() argument 392 order_code, cpu_addr); in handle_sigp_order_in_user_space() 402 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp() local 411 if (handle_sigp_order_in_user_space(vcpu, order_code, cpu_addr)) in kvm_s390_handle_sigp() 419 trace_kvm_s390_handle_sigp(vcpu, order_code, cpu_addr, parameter); in kvm_s390_handle_sigp() 427 rc = handle_sigp_dst(vcpu, order_code, cpu_addr, in kvm_s390_handle_sigp() 451 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp_pei() local 455 trace_kvm_s390_handle_sigp_pei(vcpu, order_code, cpu_addr); in kvm_s390_handle_sigp_pei() [all …]
|
/linux/arch/arm/mm/ |
A D | dma-mapping.c | 54 void *cpu_addr; member 560 dma_common_free_remap(cpu_addr, size); in __free_from_contiguous() 812 .cpu_addr = cpu_addr, in __arm_dma_free() 817 buf = arm_dma_buffer_find(cpu_addr); in __arm_dma_free() 1381 return __atomic_get_pages(cpu_addr); in __iommu_get_pages() 1384 return cpu_addr; in __iommu_get_pages() 1386 return dma_common_find_pages(cpu_addr); in __iommu_get_pages() 1421 __free_from_pool(cpu_addr, size); in __iommu_free_atomic() 1543 dma_common_free_remap(cpu_addr, size); in __arm_iommu_free_attrs() 1550 void *cpu_addr, dma_addr_t handle, in arm_iommu_free_attrs() argument [all …]
|
/linux/drivers/iommu/ |
A D | dma-iommu.c | 1130 page = virt_to_page(cpu_addr); in __iommu_dma_free() 1153 void *cpu_addr; in iommu_dma_alloc_pages() local 1166 if (!cpu_addr) in iommu_dma_alloc_pages() 1172 cpu_addr = page_address(page); in iommu_dma_alloc_pages() 1176 memset(cpu_addr, 0, alloc_size); in iommu_dma_alloc_pages() 1177 return cpu_addr; in iommu_dma_alloc_pages() 1189 void *cpu_addr; in iommu_dma_alloc() local 1205 if (!cpu_addr) in iommu_dma_alloc() 1215 return cpu_addr; in iommu_dma_alloc() 1239 pfn = vmalloc_to_pfn(cpu_addr); in iommu_dma_mmap() [all …]
|
/linux/arch/openrisc/kernel/ |
A D | dma.c | 68 void *arch_dma_set_uncached(void *cpu_addr, size_t size) in arch_dma_set_uncached() argument 70 unsigned long va = (unsigned long)cpu_addr; in arch_dma_set_uncached() 84 return cpu_addr; in arch_dma_set_uncached() 87 void arch_dma_clear_uncached(void *cpu_addr, size_t size) in arch_dma_clear_uncached() argument 89 unsigned long va = (unsigned long)cpu_addr; in arch_dma_clear_uncached()
|
/linux/drivers/media/common/saa7146/ |
A D | saa7146_core.c | 415 dev->d_rps0.cpu_addr = dma_alloc_coherent(&pci->dev, SAA7146_RPS_MEM, in saa7146_init_one() 418 if (!dev->d_rps0.cpu_addr) in saa7146_init_one() 424 if (!dev->d_rps1.cpu_addr) in saa7146_init_one() 427 dev->d_i2c.cpu_addr = dma_alloc_coherent(&pci->dev, SAA7146_RPS_MEM, in saa7146_init_one() 429 if (!dev->d_i2c.cpu_addr) in saa7146_init_one() 476 dma_free_coherent(&pci->dev, SAA7146_RPS_MEM, dev->d_i2c.cpu_addr, in saa7146_init_one() 479 dma_free_coherent(&pci->dev, SAA7146_RPS_MEM, dev->d_rps1.cpu_addr, in saa7146_init_one() 482 dma_free_coherent(&pci->dev, SAA7146_RPS_MEM, dev->d_rps0.cpu_addr, in saa7146_init_one() 505 { dev->d_i2c.cpu_addr, dev->d_i2c.dma_handle }, in saa7146_remove_one() 506 { dev->d_rps1.cpu_addr, dev->d_rps1.dma_handle }, in saa7146_remove_one() [all …]
|
/linux/arch/alpha/kernel/ |
A D | pci_iommu.c | 242 paddr = __pa(cpu_addr); in pci_map_single_1() 251 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1() 296 ret += (unsigned long)cpu_addr & ~PAGE_MASK; in pci_map_single_1() 418 void *cpu_addr; in alpha_pci_alloc_coherent() local 425 if (! cpu_addr) { in alpha_pci_alloc_coherent() 433 memset(cpu_addr, 0, size); in alpha_pci_alloc_coherent() 435 *dma_addrp = pci_map_single_1(pdev, cpu_addr, size, 0); in alpha_pci_alloc_coherent() 437 free_pages((unsigned long)cpu_addr, order); in alpha_pci_alloc_coherent() 449 return cpu_addr; in alpha_pci_alloc_coherent() 459 void *cpu_addr, dma_addr_t dma_addr, in alpha_pci_free_coherent() argument [all …]
|
/linux/drivers/pci/controller/mobiveil/ |
A D | pcie-mobiveil.c | 137 u64 cpu_addr, u64 pci_addr, u32 type, u64 size) in program_ib_windows() argument 157 mobiveil_csr_writel(pcie, lower_32_bits(cpu_addr), in program_ib_windows() 159 mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), in program_ib_windows() 174 u64 cpu_addr, u64 pci_addr, u32 type, u64 size) in program_ob_windows() argument 203 lower_32_bits(cpu_addr) & (~AXI_WINDOW_ALIGN_MASK), in program_ob_windows() 205 mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), in program_ob_windows()
|
/linux/drivers/of/ |
A D | address.c | 241 port = pci_address_to_pio(range->cpu_addr); in of_pci_range_to_resource() 249 upper_32_bits(range->cpu_addr)) { in of_pci_range_to_resource() 254 res->start = range->cpu_addr; in of_pci_range_to_resource() 744 u64 bus_addr, cpu_addr, size; in of_pci_range_parser_one() local 752 cpu_addr = of_translate_address(parser->node, in of_pci_range_parser_one() 759 cpu_addr != range->cpu_addr + range->size) in of_pci_range_parser_one() 983 if (range.cpu_addr == OF_BAD_ADDR) { in of_dma_get_range() 988 r->cpu_start = range.cpu_addr; in of_dma_get_range() 991 r->offset = range.cpu_addr - range.bus_addr; in of_dma_get_range() 1026 if (range.cpu_addr + range.size > cpu_end) in of_dma_get_max_cpu_address() [all …]
|
/linux/include/linux/ |
A D | dma-map-ops.h | 34 void *cpu_addr, dma_addr_t dma_addr, size_t size, 173 void *cpu_addr, size_t size, int *ret); 189 int dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *cpu_addr, 203 void *cpu_addr, size_t size, int *ret) in dma_mmap_from_global_coherent() argument 224 void *cpu_addr, dma_addr_t dma_addr, size_t size, 227 void *cpu_addr, dma_addr_t dma_addr, size_t size, 234 struct page **dma_common_find_pages(void *cpu_addr); 239 void dma_common_free_remap(void *cpu_addr, size_t size); 242 void **cpu_addr, gfp_t flags, 266 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
|
A D | dma-mapping.h | 129 void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, 136 void *cpu_addr, dma_addr_t dma_addr, size_t size, 139 void *cpu_addr, dma_addr_t dma_addr, size_t size, 234 struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, in dma_get_sgtable_attrs() argument 240 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_mmap_attrs() argument 423 void *cpu_addr, dma_addr_t dma_handle) in dma_free_coherent() argument 425 return dma_free_attrs(dev, size, cpu_addr, dma_handle, 0); in dma_free_coherent() 568 void *cpu_addr, dma_addr_t dma_addr) in dma_free_wc() argument 570 return dma_free_attrs(dev, size, cpu_addr, dma_addr, in dma_free_wc() 576 void *cpu_addr, dma_addr_t dma_addr, in dma_mmap_wc() argument [all …]
|
/linux/arch/mips/mm/ |
A D | ioremap.c | 54 void __iomem *cpu_addr; in ioremap_prot() local 56 cpu_addr = plat_ioremap(phys_addr, size, flags); in ioremap_prot() 57 if (cpu_addr) in ioremap_prot() 58 return cpu_addr; in ioremap_prot()
|
/linux/arch/mips/loongson64/ |
A D | init.c | 190 range.cpu_addr, in reserve_pio_range() 191 range.cpu_addr + range.size - 1, in reserve_pio_range() 193 if (add_legacy_isa_io(&np->fwnode, range.cpu_addr, range.size)) in reserve_pio_range() 198 range.cpu_addr, in reserve_pio_range() 199 range.cpu_addr + range.size - 1, in reserve_pio_range()
|
/linux/drivers/remoteproc/ |
A D | ti_k3_dsp_remoteproc.c | 34 void __iomem *cpu_addr; member 379 va = kproc->mem[i].cpu_addr + offset; in k3_dsp_rproc_da_to_va() 387 va = kproc->mem[i].cpu_addr + offset; in k3_dsp_rproc_da_to_va() 400 va = kproc->rmem[i].cpu_addr + offset; in k3_dsp_rproc_da_to_va() 446 kproc->mem[i].cpu_addr = devm_ioremap_wc(dev, res->start, in k3_dsp_rproc_of_get_memories() 448 if (!kproc->mem[i].cpu_addr) { in k3_dsp_rproc_of_get_memories() 459 kproc->mem[i].size, kproc->mem[i].cpu_addr, in k3_dsp_rproc_of_get_memories() 525 if (!kproc->rmem[i].cpu_addr) { in k3_dsp_reserved_mem_init() 534 kproc->rmem[i].size, kproc->rmem[i].cpu_addr, in k3_dsp_reserved_mem_init() 543 iounmap(kproc->rmem[i].cpu_addr); in k3_dsp_reserved_mem_init() [all …]
|
/linux/drivers/dma/ |
A D | st_fdma.h | 185 readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ 190 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ 197 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ 207 readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \ 212 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
|
/linux/drivers/gpu/drm/amd/amdgpu/ |
A D | amdgpu_fence.c | 104 if (drv->cpu_addr) in amdgpu_fence_write() 105 *drv->cpu_addr = cpu_to_le32(seq); in amdgpu_fence_write() 121 if (drv->cpu_addr) in amdgpu_fence_read() 122 seq = le32_to_cpu(*drv->cpu_addr); in amdgpu_fence_read() 425 ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs]; in amdgpu_fence_driver_start_ring() 430 ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index; in amdgpu_fence_driver_start_ring() 469 ring->fence_drv.cpu_addr = NULL; in amdgpu_fence_driver_init_ring() 806 le32_to_cpu(*(ring->fence_drv.cpu_addr + 2))); in amdgpu_debugfs_fence_info_show() 809 le32_to_cpu(*(ring->fence_drv.cpu_addr + 4))); in amdgpu_debugfs_fence_info_show() 812 le32_to_cpu(*(ring->fence_drv.cpu_addr + 6))); in amdgpu_debugfs_fence_info_show()
|
/linux/arch/microblaze/pci/ |
A D | pci-common.c | 448 range.cpu_addr, range.cpu_addr + range.size - 1, in pci_process_bridge_OF_ranges() 461 hose->io_base_virt = ioremap(range.cpu_addr, in pci_process_bridge_OF_ranges() 476 range.cpu_addr = range.pci_addr; in pci_process_bridge_OF_ranges() 481 range.cpu_addr, range.cpu_addr + range.size - 1, in pci_process_bridge_OF_ranges() 493 isa_mb = range.cpu_addr; in pci_process_bridge_OF_ranges() 496 isa_mem_base = range.cpu_addr; in pci_process_bridge_OF_ranges() 497 hose->isa_mem_phys = range.cpu_addr; in pci_process_bridge_OF_ranges() 508 hose->pci_mem_offset = range.cpu_addr - in pci_process_bridge_OF_ranges() 511 hose->pci_mem_offset != range.cpu_addr - in pci_process_bridge_OF_ranges() 524 res->start = range.cpu_addr; in pci_process_bridge_OF_ranges() [all …]
|