Searched refs:pages_mapped (Results 1 – 4 of 4) sorted by relevance
/linux-6.6.21/arch/parisc/kernel/ |
D | pci-dma.c | 309 unsigned int pages_mapped = size >> PAGE_SHIFT; in pcxl_free_range() local 312 mask >>= BITS_PER_LONG - pages_mapped; in pcxl_free_range() 315 res_idx, size, pages_mapped, mask); in pcxl_free_range() 319 if(pages_mapped <= 8) { in pcxl_free_range() 321 } else if(pages_mapped <= 16) { in pcxl_free_range() 323 } else if(pages_mapped <= 32) { in pcxl_free_range() 330 pcxl_used_pages -= (pages_mapped ? pages_mapped : 1); in pcxl_free_range() 331 pcxl_used_bytes -= ((pages_mapped >> 3) ? (pages_mapped >> 3) : 1); in pcxl_free_range()
|
/linux-6.6.21/drivers/parisc/ |
D | ccio-dma.c | 421 ccio_free_range(struct ioc *ioc, dma_addr_t iova, unsigned long pages_mapped) in ccio_free_range() argument 426 BUG_ON(pages_mapped == 0); in ccio_free_range() 427 BUG_ON((pages_mapped * IOVP_SIZE) > DMA_CHUNK_SIZE); in ccio_free_range() 428 BUG_ON(pages_mapped > BITS_PER_LONG); in ccio_free_range() 431 __func__, res_idx, pages_mapped); in ccio_free_range() 434 ioc->used_pages -= pages_mapped; in ccio_free_range() 437 if(pages_mapped <= 8) { in ccio_free_range() 440 unsigned long mask = ~(~0UL >> pages_mapped); in ccio_free_range() 445 } else if(pages_mapped <= 16) { in ccio_free_range() 447 } else if(pages_mapped <= 32) { in ccio_free_range() [all …]
|
/linux-6.6.21/drivers/infiniband/hw/mlx5/ |
D | umr.c | 699 size_t pages_mapped = 0; in mlx5r_umr_update_xlt() local 742 for (pages_mapped = 0; in mlx5r_umr_update_xlt() 743 pages_mapped < pages_to_map && !err; in mlx5r_umr_update_xlt() 744 pages_mapped += pages_iter, idx += pages_iter) { in mlx5r_umr_update_xlt() 745 npages = min_t(int, pages_iter, pages_to_map - pages_mapped); in mlx5r_umr_update_xlt() 754 if (pages_mapped + pages_iter >= pages_to_map) in mlx5r_umr_update_xlt()
|
/linux-6.6.21/net/ipv4/ |
D | tcp.c | 2004 unsigned int pages_mapped; in tcp_zerocopy_vm_insert_batch() local 2009 pages_mapped = pages_to_map - (unsigned int)pages_remaining; in tcp_zerocopy_vm_insert_batch() 2010 bytes_mapped = PAGE_SIZE * pages_mapped; in tcp_zerocopy_vm_insert_batch() 2021 return tcp_zerocopy_vm_insert_batch_error(vma, pages + pages_mapped, in tcp_zerocopy_vm_insert_batch()
|