Home
last modified time | relevance | path

Searched refs:cpu_addr (Results 1 – 25 of 174) sorted by relevance

1234567

/linux-6.1.9/include/trace/events/
Dhabanalabs.h54 TP_PROTO(struct device *dev, u64 cpu_addr, u64 dma_addr, size_t size, const char *caller),
56 TP_ARGS(dev, cpu_addr, dma_addr, size, caller),
60 __field(u64, cpu_addr)
68 __entry->cpu_addr = cpu_addr;
76 __entry->cpu_addr,
83 TP_PROTO(struct device *dev, u64 cpu_addr, u64 dma_addr, size_t size, const char *caller),
84 TP_ARGS(dev, cpu_addr, dma_addr, size, caller));
87 TP_PROTO(struct device *dev, u64 cpu_addr, u64 dma_addr, size_t size, const char *caller),
88 TP_ARGS(dev, cpu_addr, dma_addr, size, caller));
/linux-6.1.9/kernel/dma/
Dops_helpers.c8 static struct page *dma_common_vaddr_to_page(void *cpu_addr) in dma_common_vaddr_to_page() argument
10 if (is_vmalloc_addr(cpu_addr)) in dma_common_vaddr_to_page()
11 return vmalloc_to_page(cpu_addr); in dma_common_vaddr_to_page()
12 return virt_to_page(cpu_addr); in dma_common_vaddr_to_page()
19 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_common_get_sgtable() argument
22 struct page *page = dma_common_vaddr_to_page(cpu_addr); in dma_common_get_sgtable()
35 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_common_mmap() argument
42 struct page *page = dma_common_vaddr_to_page(cpu_addr); in dma_common_mmap()
47 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in dma_common_mmap()
Dremap.c9 struct page **dma_common_find_pages(void *cpu_addr) in dma_common_find_pages() argument
11 struct vm_struct *area = find_vm_area(cpu_addr); in dma_common_find_pages()
60 void dma_common_free_remap(void *cpu_addr, size_t size) in dma_common_free_remap() argument
62 struct vm_struct *area = find_vm_area(cpu_addr); in dma_common_free_remap()
65 WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); in dma_common_free_remap()
69 vunmap(cpu_addr); in dma_common_free_remap()
Dmapping.c396 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_get_sgtable_attrs() argument
402 return dma_direct_get_sgtable(dev, sgt, cpu_addr, dma_addr, in dma_get_sgtable_attrs()
406 return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs); in dma_get_sgtable_attrs()
458 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_mmap_attrs() argument
464 return dma_direct_mmap(dev, vma, cpu_addr, dma_addr, size, in dma_mmap_attrs()
468 return ops->mmap(dev, vma, cpu_addr, dma_addr, size, attrs); in dma_mmap_attrs()
497 void *cpu_addr; in dma_alloc_attrs() local
501 if (dma_alloc_from_dev_coherent(dev, size, dma_handle, &cpu_addr)) in dma_alloc_attrs()
502 return cpu_addr; in dma_alloc_attrs()
508 cpu_addr = dma_direct_alloc(dev, size, dma_handle, flag, attrs); in dma_alloc_attrs()
[all …]
Ddirect.c323 void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs) in dma_direct_free() argument
330 dma_free_contiguous(dev, cpu_addr, size); in dma_direct_free()
339 arch_dma_free(dev, size, cpu_addr, dma_addr, attrs); in dma_direct_free()
345 if (!dma_release_from_global_coherent(page_order, cpu_addr)) in dma_direct_free()
352 dma_free_from_pool(dev, cpu_addr, PAGE_ALIGN(size))) in dma_direct_free()
355 if (is_vmalloc_addr(cpu_addr)) { in dma_direct_free()
356 vunmap(cpu_addr); in dma_direct_free()
359 arch_dma_clear_uncached(cpu_addr, size); in dma_direct_free()
360 if (dma_set_encrypted(dev, cpu_addr, size)) in dma_direct_free()
537 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_direct_get_sgtable() argument
[all …]
/linux-6.1.9/drivers/pci/controller/cadence/
Dpcie-cadence.c28 u64 cpu_addr, u64 pci_addr, size_t size) in cdns_pcie_set_outbound_region() argument
93 cpu_addr = pcie->ops->cpu_addr_fixup(pcie, cpu_addr); in cdns_pcie_set_outbound_region()
96 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_set_outbound_region()
97 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_set_outbound_region()
105 u32 r, u64 cpu_addr) in cdns_pcie_set_outbound_region_for_normal_msg() argument
123 cpu_addr = pcie->ops->cpu_addr_fixup(pcie, cpu_addr); in cdns_pcie_set_outbound_region_for_normal_msg()
126 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_set_outbound_region_for_normal_msg()
127 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_set_outbound_region_for_normal_msg()
Dpcie-cadence-host.c193 u64 cpu_addr, u64 size, in cdns_pcie_host_bar_ib_config() argument
206 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_host_bar_ib_config()
207 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_host_bar_ib_config()
220 if (size + cpu_addr >= SZ_4G) { in cdns_pcie_host_bar_ib_config()
287 u64 cpu_addr, pci_addr, size, winsize; in cdns_pcie_host_bar_config() local
294 cpu_addr = entry->res->start; in cdns_pcie_host_bar_config()
301 pci_addr, cpu_addr); in cdns_pcie_host_bar_config()
316 ret = cdns_pcie_host_bar_ib_config(rc, bar, cpu_addr, in cdns_pcie_host_bar_config()
337 cpu_addr); in cdns_pcie_host_bar_config()
342 ret = cdns_pcie_host_bar_ib_config(rc, bar, cpu_addr, winsize, in cdns_pcie_host_bar_config()
[all …]
/linux-6.1.9/arch/s390/kvm/
Dtrace.h220 TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr, \
222 TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr, parameter),
227 __field(__u16, cpu_addr)
234 __entry->cpu_addr = cpu_addr;
242 __entry->cpu_addr, __entry->parameter)
246 TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr),
247 TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr),
252 __field(__u16, cpu_addr)
258 __entry->cpu_addr = cpu_addr;
265 __entry->cpu_addr)
Dsigp.c271 u16 cpu_addr, u32 parameter, u64 *status_reg) in handle_sigp_dst() argument
274 struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr); in handle_sigp_dst()
376 u16 cpu_addr) in handle_sigp_order_in_user_space() argument
420 order_code, cpu_addr); in handle_sigp_order_in_user_space()
430 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp() local
439 if (handle_sigp_order_in_user_space(vcpu, order_code, cpu_addr)) in kvm_s390_handle_sigp()
447 trace_kvm_s390_handle_sigp(vcpu, order_code, cpu_addr, parameter); in kvm_s390_handle_sigp()
455 rc = handle_sigp_dst(vcpu, order_code, cpu_addr, in kvm_s390_handle_sigp()
479 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp_pei() local
484 trace_kvm_s390_handle_sigp_pei(vcpu, order_code, cpu_addr); in kvm_s390_handle_sigp_pei()
[all …]
/linux-6.1.9/arch/openrisc/kernel/
Ddma.c68 void *arch_dma_set_uncached(void *cpu_addr, size_t size) in arch_dma_set_uncached() argument
70 unsigned long va = (unsigned long)cpu_addr; in arch_dma_set_uncached()
84 return cpu_addr; in arch_dma_set_uncached()
87 void arch_dma_clear_uncached(void *cpu_addr, size_t size) in arch_dma_clear_uncached() argument
89 unsigned long va = (unsigned long)cpu_addr; in arch_dma_clear_uncached()
/linux-6.1.9/drivers/pci/controller/dwc/
Dpcie-designware.c310 int index, int type, u64 cpu_addr, in __dw_pcie_prog_outbound_atu() argument
317 cpu_addr = pci->ops->cpu_addr_fixup(pci, cpu_addr); in __dw_pcie_prog_outbound_atu()
319 limit_addr = cpu_addr + size - 1; in __dw_pcie_prog_outbound_atu()
321 if ((limit_addr & ~pci->region_limit) != (cpu_addr & ~pci->region_limit) || in __dw_pcie_prog_outbound_atu()
322 !IS_ALIGNED(cpu_addr, pci->region_align) || in __dw_pcie_prog_outbound_atu()
328 lower_32_bits(cpu_addr)); in __dw_pcie_prog_outbound_atu()
330 upper_32_bits(cpu_addr)); in __dw_pcie_prog_outbound_atu()
344 if (upper_32_bits(limit_addr) > upper_32_bits(cpu_addr) && in __dw_pcie_prog_outbound_atu()
371 u64 cpu_addr, u64 pci_addr, u64 size) in dw_pcie_prog_outbound_atu() argument
374 cpu_addr, pci_addr, size); in dw_pcie_prog_outbound_atu()
[all …]
Dpcie-tegra194-acpi.c49 int index, int type, u64 cpu_addr, in program_outbound_atu() argument
52 atu_reg_write(pcie_ecam, index, lower_32_bits(cpu_addr), in program_outbound_atu()
54 atu_reg_write(pcie_ecam, index, upper_32_bits(cpu_addr), in program_outbound_atu()
58 atu_reg_write(pcie_ecam, index, lower_32_bits(cpu_addr + size - 1), in program_outbound_atu()
/linux-6.1.9/drivers/pci/controller/mobiveil/
Dpcie-mobiveil.c137 u64 cpu_addr, u64 pci_addr, u32 type, u64 size) in program_ib_windows() argument
157 mobiveil_csr_writel(pcie, lower_32_bits(cpu_addr), in program_ib_windows()
159 mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), in program_ib_windows()
174 u64 cpu_addr, u64 pci_addr, u32 type, u64 size) in program_ob_windows() argument
203 lower_32_bits(cpu_addr) & (~AXI_WINDOW_ALIGN_MASK), in program_ob_windows()
205 mobiveil_csr_writel(pcie, upper_32_bits(cpu_addr), in program_ob_windows()
/linux-6.1.9/arch/mips/mm/
Dioremap.c54 void __iomem *cpu_addr; in ioremap_prot() local
56 cpu_addr = plat_ioremap(phys_addr, size, flags); in ioremap_prot()
57 if (cpu_addr) in ioremap_prot()
58 return cpu_addr; in ioremap_prot()
/linux-6.1.9/drivers/staging/media/deprecated/saa7146/common/
Dsaa7146_core.c415 dev->d_rps0.cpu_addr = dma_alloc_coherent(&pci->dev, SAA7146_RPS_MEM, in saa7146_init_one()
418 if (!dev->d_rps0.cpu_addr) in saa7146_init_one()
421 dev->d_rps1.cpu_addr = dma_alloc_coherent(&pci->dev, SAA7146_RPS_MEM, in saa7146_init_one()
424 if (!dev->d_rps1.cpu_addr) in saa7146_init_one()
427 dev->d_i2c.cpu_addr = dma_alloc_coherent(&pci->dev, SAA7146_RPS_MEM, in saa7146_init_one()
429 if (!dev->d_i2c.cpu_addr) in saa7146_init_one()
476 dma_free_coherent(&pci->dev, SAA7146_RPS_MEM, dev->d_i2c.cpu_addr, in saa7146_init_one()
479 dma_free_coherent(&pci->dev, SAA7146_RPS_MEM, dev->d_rps1.cpu_addr, in saa7146_init_one()
482 dma_free_coherent(&pci->dev, SAA7146_RPS_MEM, dev->d_rps0.cpu_addr, in saa7146_init_one()
505 { dev->d_i2c.cpu_addr, dev->d_i2c.dma_handle }, in saa7146_remove_one()
[all …]
/linux-6.1.9/include/linux/
Ddma-map-ops.h44 void *cpu_addr, dma_addr_t dma_addr, size_t size,
185 void *cpu_addr, size_t size, int *ret);
203 int dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *cpu_addr,
217 void *cpu_addr, size_t size, int *ret) in dma_mmap_from_global_coherent() argument
238 void *cpu_addr, dma_addr_t dma_addr, size_t size,
241 void *cpu_addr, dma_addr_t dma_addr, size_t size,
248 struct page **dma_common_find_pages(void *cpu_addr);
253 void dma_common_free_remap(void *cpu_addr, size_t size);
256 void **cpu_addr, gfp_t flags,
280 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
Ddma-mapping.h129 void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr,
136 void *cpu_addr, dma_addr_t dma_addr, size_t size,
139 void *cpu_addr, dma_addr_t dma_addr, size_t size,
221 static void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, in dma_free_attrs() argument
235 struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, in dma_get_sgtable_attrs() argument
241 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_mmap_attrs() argument
428 void *cpu_addr, dma_addr_t dma_handle) in dma_free_coherent() argument
430 return dma_free_attrs(dev, size, cpu_addr, dma_handle, 0); in dma_free_coherent()
573 void *cpu_addr, dma_addr_t dma_addr) in dma_free_wc() argument
575 return dma_free_attrs(dev, size, cpu_addr, dma_addr, in dma_free_wc()
[all …]
/linux-6.1.9/arch/alpha/kernel/
Dpci_iommu.c230 pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size, in pci_map_single_1() argument
242 paddr = __pa(cpu_addr); in pci_map_single_1()
251 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1()
262 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1()
296 ret += (unsigned long)cpu_addr & ~PAGE_MASK; in pci_map_single_1()
299 cpu_addr, size, npages, ret, __builtin_return_address(0)); in pci_map_single_1()
418 void *cpu_addr; in alpha_pci_alloc_coherent() local
424 cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order); in alpha_pci_alloc_coherent()
425 if (! cpu_addr) { in alpha_pci_alloc_coherent()
433 memset(cpu_addr, 0, size); in alpha_pci_alloc_coherent()
[all …]
/linux-6.1.9/drivers/iommu/
Ddma-iommu.c1367 static void __iommu_dma_free(struct device *dev, size_t size, void *cpu_addr) in __iommu_dma_free() argument
1375 dma_free_from_pool(dev, cpu_addr, alloc_size)) in __iommu_dma_free()
1378 if (is_vmalloc_addr(cpu_addr)) { in __iommu_dma_free()
1383 pages = dma_common_find_pages(cpu_addr); in __iommu_dma_free()
1385 page = vmalloc_to_page(cpu_addr); in __iommu_dma_free()
1386 dma_common_free_remap(cpu_addr, alloc_size); in __iommu_dma_free()
1389 page = virt_to_page(cpu_addr); in __iommu_dma_free()
1398 static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr, in iommu_dma_free() argument
1402 __iommu_dma_free(dev, size, cpu_addr); in iommu_dma_free()
1412 void *cpu_addr; in iommu_dma_alloc_pages() local
[all …]
/linux-6.1.9/drivers/of/
Daddress.c237 err = pci_register_io_range(&np->fwnode, range->cpu_addr, in of_pci_range_to_resource()
241 port = pci_address_to_pio(range->cpu_addr); in of_pci_range_to_resource()
249 upper_32_bits(range->cpu_addr)) { in of_pci_range_to_resource()
254 res->start = range->cpu_addr; in of_pci_range_to_resource()
734 range->cpu_addr = of_translate_dma_address(parser->node, in of_pci_range_parser_one()
737 range->cpu_addr = of_translate_address(parser->node, in of_pci_range_parser_one()
746 u64 bus_addr, cpu_addr, size; in of_pci_range_parser_one() local
751 cpu_addr = of_translate_dma_address(parser->node, in of_pci_range_parser_one()
754 cpu_addr = of_translate_address(parser->node, in of_pci_range_parser_one()
761 cpu_addr != range->cpu_addr + range->size) in of_pci_range_parser_one()
[all …]
/linux-6.1.9/arch/mips/loongson64/
Dinit.c190 range.cpu_addr, in reserve_pio_range()
191 range.cpu_addr + range.size - 1, in reserve_pio_range()
193 if (add_legacy_isa_io(&np->fwnode, range.cpu_addr, range.size)) in reserve_pio_range()
198 range.cpu_addr, in reserve_pio_range()
199 range.cpu_addr + range.size - 1, in reserve_pio_range()
/linux-6.1.9/arch/arm/mm/
Ddma-mapping.c54 void *cpu_addr; member
425 void *cpu_addr, size_t size, bool want_vaddr) in __free_from_contiguous() argument
429 dma_common_free_remap(cpu_addr, size); in __free_from_contiguous()
485 __free_from_contiguous(args->dev, args->page, args->cpu_addr, in cma_allocator_free()
502 __free_from_pool(args->cpu_addr, args->size); in pool_allocator_free()
521 dma_common_free_remap(args->cpu_addr, args->size); in remap_allocator_free()
611 static void __arm_dma_free(struct device *dev, size_t size, void *cpu_addr, in __arm_dma_free() argument
620 .cpu_addr = cpu_addr, in __arm_dma_free()
625 buf = arm_dma_buffer_find(cpu_addr); in __arm_dma_free()
626 if (WARN(!buf, "Freeing invalid buffer %p\n", cpu_addr)) in __arm_dma_free()
[all …]
/linux-6.1.9/drivers/dma/
Dst_fdma.h185 readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
190 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
197 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
207 readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
212 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
/linux-6.1.9/drivers/remoteproc/
Dti_k3_dsp_remoteproc.c34 void __iomem *cpu_addr; member
419 if (!kproc->rmem[0].cpu_addr) { in k3_dsp_get_loaded_rsc_table()
432 return (struct resource_table *)kproc->rmem[0].cpu_addr; in k3_dsp_get_loaded_rsc_table()
465 va = kproc->mem[i].cpu_addr + offset; in k3_dsp_rproc_da_to_va()
473 va = kproc->mem[i].cpu_addr + offset; in k3_dsp_rproc_da_to_va()
486 va = kproc->rmem[i].cpu_addr + offset; in k3_dsp_rproc_da_to_va()
532 kproc->mem[i].cpu_addr = devm_ioremap_wc(dev, res->start, in k3_dsp_rproc_of_get_memories()
534 if (!kproc->mem[i].cpu_addr) { in k3_dsp_rproc_of_get_memories()
545 kproc->mem[i].size, kproc->mem[i].cpu_addr, in k3_dsp_rproc_of_get_memories()
610 kproc->rmem[i].cpu_addr = ioremap_wc(rmem->base, rmem->size); in k3_dsp_reserved_mem_init()
[all …]
/linux-6.1.9/drivers/gpu/drm/amd/amdgpu/
Damdgpu_fence.c105 if (drv->cpu_addr) in amdgpu_fence_write()
106 *drv->cpu_addr = cpu_to_le32(seq); in amdgpu_fence_write()
122 if (drv->cpu_addr) in amdgpu_fence_read()
123 seq = le32_to_cpu(*drv->cpu_addr); in amdgpu_fence_read()
430 ring->fence_drv.cpu_addr = ring->fence_cpu_addr; in amdgpu_fence_driver_start_ring()
435 ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index; in amdgpu_fence_driver_start_ring()
468 ring->fence_drv.cpu_addr = NULL; in amdgpu_fence_driver_init_ring()
798 le32_to_cpu(*(ring->fence_drv.cpu_addr + 2))); in amdgpu_debugfs_fence_info_show()
801 le32_to_cpu(*(ring->fence_drv.cpu_addr + 4))); in amdgpu_debugfs_fence_info_show()
804 le32_to_cpu(*(ring->fence_drv.cpu_addr + 6))); in amdgpu_debugfs_fence_info_show()

1234567