/linux-6.6.21/include/linux/ |
D | iommu.h | 456 static inline void iommu_iotlb_gather_init(struct iommu_iotlb_gather *gather) in iommu_iotlb_gather_init() argument 458 *gather = (struct iommu_iotlb_gather) { in iommu_iotlb_gather_init() 460 .freelist = LIST_HEAD_INIT(gather->freelist), in iommu_iotlb_gather_init() 571 bool iommu_iotlb_gather_is_disjoint(struct iommu_iotlb_gather *gather, in iommu_iotlb_gather_is_disjoint() argument 576 return gather->end != 0 && in iommu_iotlb_gather_is_disjoint() 577 (end + 1 < gather->start || start > gather->end + 1); in iommu_iotlb_gather_is_disjoint() 591 static inline void iommu_iotlb_gather_add_range(struct iommu_iotlb_gather *gather, in iommu_iotlb_gather_add_range() argument 596 if (gather->start > iova) in iommu_iotlb_gather_add_range() 597 gather->start = iova; in iommu_iotlb_gather_add_range() 598 if (gather->end < end) in iommu_iotlb_gather_add_range() [all …]
|
D | io-pgtable.h | 45 void (*tlb_add_page)(struct iommu_iotlb_gather *gather, 166 struct iommu_iotlb_gather *gather); 233 struct iommu_iotlb_gather * gather, unsigned long iova, in io_pgtable_tlb_add_page() argument 237 iop->cfg.tlb->tlb_add_page(gather, iova, granule, iop->cookie); in io_pgtable_tlb_add_page()
|
/linux-6.6.21/drivers/gpu/host1x/ |
D | job.c | 116 struct host1x_job_gather *gather = &job->cmds[job->num_cmds].gather; in host1x_job_add_gather() local 118 gather->words = words; in host1x_job_add_gather() 119 gather->bo = bo; in host1x_job_add_gather() 120 gather->offset = offset; in host1x_job_add_gather() 223 g = &job->cmds[i].gather; in pin_job() 544 g = &job->cmds[i].gather; in copy_gathers() 568 void *gather; in copy_gathers() local 572 g = &job->cmds[i].gather; in copy_gathers() 575 gather = host1x_bo_mmap(g->bo); in copy_gathers() 576 memcpy(job->gather_copy_mapped + offset, gather + g->offset, in copy_gathers() [all …]
|
D | job.h | 32 struct host1x_job_gather gather; member
|
/linux-6.6.21/drivers/net/ethernet/cavium/liquidio/ |
D | octeon_nic.h | 95 u32 gather:1; member 150 if (!setup->s.gather) { in octnet_prepare_pci_cmd_o2() 153 ih2->gather = 1; in octnet_prepare_pci_cmd_o2() 195 if (!setup->s.gather) { in octnet_prepare_pci_cmd_o3() 198 ih3->gather = 1; in octnet_prepare_pci_cmd_o3()
|
D | liquidio_common.h | 357 u64 gather:1; member 391 u64 gather:1; 488 u64 gather:1; member 533 u64 gather:1;
|
/linux-6.6.21/drivers/iommu/ |
D | s390-iommu.c | 222 struct iommu_iotlb_gather *gather) in s390_iommu_iotlb_sync() argument 225 size_t size = gather->end - gather->start + 1; in s390_iommu_iotlb_sync() 229 if (!gather->end) in s390_iommu_iotlb_sync() 234 zpci_refresh_trans((u64)zdev->fh << 32, gather->start, in s390_iommu_iotlb_sync() 383 struct iommu_iotlb_gather *gather) in s390_iommu_unmap_pages() argument 397 iommu_iotlb_gather_add_range(gather, iova, size); in s390_iommu_unmap_pages()
|
D | tegra-gart.c | 210 size_t bytes, struct iommu_iotlb_gather *gather) in gart_iommu_unmap() argument 262 struct iommu_iotlb_gather *gather) in gart_iommu_sync() argument 264 size_t length = gather->end - gather->start + 1; in gart_iommu_sync() 266 gart_iommu_sync_map(domain, gather->start, length); in gart_iommu_sync()
|
D | io-pgtable-arm-v7s.c | 622 struct iommu_iotlb_gather *gather, in arm_v7s_split_blk_unmap() argument 659 return __arm_v7s_unmap(data, gather, iova, size, 2, tablep); in arm_v7s_split_blk_unmap() 662 io_pgtable_tlb_add_page(&data->iop, gather, iova, size); in arm_v7s_split_blk_unmap() 667 struct iommu_iotlb_gather *gather, in __arm_v7s_unmap() argument 718 } else if (!iommu_iotlb_gather_queued(gather)) { in __arm_v7s_unmap() 719 io_pgtable_tlb_add_page(iop, gather, iova, blk_size); in __arm_v7s_unmap() 729 return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0], in __arm_v7s_unmap() 735 return __arm_v7s_unmap(data, gather, iova, size, lvl + 1, ptep); in __arm_v7s_unmap() 740 struct iommu_iotlb_gather *gather) in arm_v7s_unmap_pages() argument 749 ret = __arm_v7s_unmap(data, gather, iova, pgsize, 1, data->pgd); in arm_v7s_unmap_pages() [all …]
|
D | io-pgtable-arm.c | 254 struct iommu_iotlb_gather *gather, 539 struct iommu_iotlb_gather *gather, in arm_lpae_split_blk_unmap() argument 590 io_pgtable_tlb_add_page(&data->iop, gather, iova + i * size, size); in arm_lpae_split_blk_unmap() 595 return __arm_lpae_unmap(data, gather, iova, size, pgcount, lvl, tablep); in arm_lpae_split_blk_unmap() 599 struct iommu_iotlb_gather *gather, in __arm_lpae_unmap() argument 634 } else if (!iommu_iotlb_gather_queued(gather)) { in __arm_lpae_unmap() 635 io_pgtable_tlb_add_page(iop, gather, iova + i * size, size); in __arm_lpae_unmap() 648 return arm_lpae_split_blk_unmap(data, gather, iova, size, pte, in __arm_lpae_unmap() 654 return __arm_lpae_unmap(data, gather, iova, size, pgcount, lvl + 1, ptep); in __arm_lpae_unmap() 659 struct iommu_iotlb_gather *gather) in arm_lpae_unmap_pages() argument [all …]
|
D | io-pgtable-dart.c | 300 struct iommu_iotlb_gather *gather) in dart_unmap_pages() argument 330 if (!iommu_iotlb_gather_queued(gather)) in dart_unmap_pages() 331 io_pgtable_tlb_add_page(&data->iop, gather, in dart_unmap_pages()
|
/linux-6.6.21/drivers/net/ethernet/marvell/octeon_ep/ |
D | octep_tx.h | 55 u8 gather; member 237 u64 gather:1; member
|
D | octep_tx.c | 59 if (!tx_buffer->gather) { in octep_iq_process_completions() 121 if (!tx_buffer->gather) { in octep_iq_free_pending()
|
/linux-6.6.21/Documentation/admin-guide/hw-vuln/ |
D | gather_data_sampling.rst | 11 When a gather instruction performs loads from memory, different data elements 12 are merged into the destination vector register. However, when a gather 19 The attacker uses gather instructions to infer the stale vector register data. 21 registers. The victim does not need to use gather instructions to be
|
/linux-6.6.21/drivers/gpu/drm/tegra/ |
D | submit.c | 70 struct gather_bo *gather = container_of(bo, struct gather_bo, base); in gather_bo_pin() local 89 err = dma_get_sgtable(gather->dev, map->sgt, gather->gather_data, gather->gather_data_dma, in gather_bo_pin() 90 gather->gather_data_words * 4); in gather_bo_pin() 99 map->size = gather->gather_data_words * 4; in gather_bo_pin()
|
/linux-6.6.21/drivers/crypto/cavium/zip/ |
D | zip_device.h | 57 union zip_zptr_s *gather; member
|
/linux-6.6.21/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 558 struct sq_gather_subdesc *gather; in nicvf_unmap_sndq_buffers() local 564 gather = (struct sq_gather_subdesc *)GET_SQ_DESC(sq, hdr_sqe); in nicvf_unmap_sndq_buffers() 566 dma_unmap_page_attrs(&nic->pdev->dev, gather->addr, in nicvf_unmap_sndq_buffers() 567 gather->size, DMA_TO_DEVICE, in nicvf_unmap_sndq_buffers() 1422 struct sq_gather_subdesc *gather; in nicvf_sq_add_gather_subdesc() local 1425 gather = (struct sq_gather_subdesc *)GET_SQ_DESC(sq, qentry); in nicvf_sq_add_gather_subdesc() 1427 memset(gather, 0, SND_QUEUE_DESC_SIZE); in nicvf_sq_add_gather_subdesc() 1428 gather->subdesc_type = SQ_DESC_TYPE_GATHER; in nicvf_sq_add_gather_subdesc() 1429 gather->ld_type = NIC_SEND_LD_TYPE_E_LDD; in nicvf_sq_add_gather_subdesc() 1430 gather->size = size; in nicvf_sq_add_gather_subdesc() [all …]
|
/linux-6.6.21/Documentation/driver-api/fpga/ |
D | fpga-programming.rst | 55 if (image is in a scatter gather table) { 57 info->sgt = [your scatter gather table]
|
D | fpga-mgr.rst | 13 The FPGA image to be programmed can be in a scatter gather list, a single 16 gather list instead if possible. 95 whole buffer will be passed into .parse_header. If image is in scatter-gather
|
/linux-6.6.21/drivers/message/fusion/ |
D | Kconfig | 65 int "Maximum number of scatter gather entries (16 - 128)" 70 gather entries per I/O. The driver default is 128, which matches
|
/linux-6.6.21/arch/mips/include/asm/octeon/ |
D | cvmx-pko.h | 220 uint64_t gather:1; member 250 uint64_t gather:1;
|
/linux-6.6.21/drivers/scsi/mpt3sas/ |
D | Kconfig | 59 gather entries per I/O. The driver default is 128, which matches 71 gather entries per I/O. The driver default is 128, which matches
|
/linux-6.6.21/drivers/iommu/amd/ |
D | io_pgtable_v2.c | 291 struct iommu_iotlb_gather *gather) in iommu_v2_unmap_pages() argument 345 static void v2_tlb_add_page(struct iommu_iotlb_gather *gather, in v2_tlb_add_page() argument
|
/linux-6.6.21/drivers/iommu/arm/arm-smmu/ |
D | qcom_iommu.c | 186 static void qcom_iommu_tlb_add_page(struct iommu_iotlb_gather *gather, in qcom_iommu_tlb_add_page() argument 423 struct iommu_iotlb_gather *gather) in qcom_iommu_unmap() argument 440 ret = ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in qcom_iommu_unmap() 461 struct iommu_iotlb_gather *gather) in qcom_iommu_iotlb_sync() argument
|
/linux-6.6.21/kernel/dma/ |
D | Kconfig | 151 hardware components that do not support I/O map nor scatter-gather. 249 bool "Debug DMA scatter-gather usage" 259 preparing literal scatter-gather descriptors, where there is a risk of
|