/linux-5.19.10/include/xen/ |
D | page.h | 9 #define XEN_PAGE_SIZE (_AC(1, UL) << XEN_PAGE_SHIFT) macro 10 #define XEN_PAGE_MASK (~(XEN_PAGE_SIZE-1)) 23 #define XEN_PFN_PER_PAGE (PAGE_SIZE / XEN_PAGE_SIZE) 26 #define XEN_PFN_UP(x) (((x) + XEN_PAGE_SIZE-1) >> XEN_PAGE_SHIFT)
|
D | grant_table.h | 318 len = min_t(unsigned int, XEN_PAGE_SIZE - (offset & ~XEN_PAGE_MASK), in gnttab_for_one_grant()
|
/linux-5.19.10/drivers/net/xen-netback/ |
D | common.h | 69 #define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, XEN_PAGE_SIZE) 70 #define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, XEN_PAGE_SIZE) 85 #define MAX_BUFFER_OFFSET XEN_PAGE_SIZE 92 #define MAX_XEN_SKB_FRAGS (65536 / XEN_PAGE_SIZE + 1)
|
D | rx.c | 48 needed = DIV_ROUND_UP(skb->len, XEN_PAGE_SIZE); in xenvif_update_needed_slots() 377 chunk_len = min_t(size_t, frag_len, XEN_PAGE_SIZE - offset); in xenvif_rx_next_chunk() 378 chunk_len = min_t(size_t, chunk_len, XEN_PAGE_SIZE - in xenvif_rx_next_chunk() 409 } while (offset < XEN_PAGE_SIZE && pkt->remaining_len > 0); in xenvif_rx_data_slot()
|
D | hash.c | 349 len > XEN_PAGE_SIZE / sizeof(*mapping)) in xenvif_set_hash_mapping() 354 if (copy_op[0].dest.offset + copy_op[0].len > XEN_PAGE_SIZE) { in xenvif_set_hash_mapping() 356 copy_op[1].source.offset = XEN_PAGE_SIZE - copy_op[0].dest.offset; in xenvif_set_hash_mapping()
|
D | netback.c | 309 if (unlikely((txp->offset + txp->size) > XEN_PAGE_SIZE)) { in xenvif_count_requests() 901 if (unlikely((txreq.offset + txreq.size) > XEN_PAGE_SIZE)) { in xenvif_tx_build_gops() 1488 BACK_RING_ATTACH(&queue->tx, txs, rsp_prod, XEN_PAGE_SIZE); in xenvif_map_frontend_data_rings() 1503 BACK_RING_ATTACH(&queue->rx, rxs, rsp_prod, XEN_PAGE_SIZE); in xenvif_map_frontend_data_rings()
|
/linux-5.19.10/drivers/xen/ |
D | mem-reservation.c | 46 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_update() 72 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in __xenmem_reservation_va_mapping_reset()
|
D | arm-device.c | 24 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_unmap_device_mmio() 58 nr = DIV_ROUND_UP(resource_size(r), XEN_PAGE_SIZE); in xen_map_device_mmio()
|
D | biomerge.c | 11 #if XEN_PAGE_SIZE == PAGE_SIZE in xen_biovec_phys_mergeable()
|
D | grant-table.c | 812 vaddr = memremap(addr, XEN_PAGE_SIZE * max_nr_gframes, MEMREMAP_WB); in gnttab_setup_auto_xlat_frames() 1197 glen = min_t(unsigned int, XEN_PAGE_SIZE - goffset, len); in gnttab_foreach_grant_in_range() 1222 fn(pfn_to_gfn(xen_pfn), goffset, XEN_PAGE_SIZE, data); in gnttab_foreach_grant() 1224 goffset += XEN_PAGE_SIZE; in gnttab_foreach_grant() 1489 .grefs_per_grant_frame = XEN_PAGE_SIZE / 1500 .grefs_per_grant_frame = XEN_PAGE_SIZE /
|
D | balloon.c | 308 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in reserve_additional_memory() 637 BUILD_BUG_ON(XEN_PAGE_SIZE != PAGE_SIZE); in xen_alloc_ballooned_pages()
|
D | gntdev.c | 862 if (seg->source.foreign.offset + seg->len > XEN_PAGE_SIZE) in gntdev_grant_copy_seg() 866 if (seg->dest.foreign.offset + seg->len > XEN_PAGE_SIZE) in gntdev_grant_copy_seg() 899 len = min(len, (size_t)XEN_PAGE_SIZE - off); in gntdev_grant_copy_seg() 919 len = min(len, (size_t)XEN_PAGE_SIZE - off); in gntdev_grant_copy_seg()
|
/linux-5.19.10/arch/arm/xen/ |
D | mm.c | 57 if (size + cflush.offset > XEN_PAGE_SIZE) in dma_cache_maint() 58 cflush.length = XEN_PAGE_SIZE - cflush.offset; in dma_cache_maint()
|
/linux-5.19.10/drivers/xen/xenbus/ |
D | xenbus_client.c | 382 unsigned long ring_size = nr_pages * XEN_PAGE_SIZE; in xenbus_setup_ring() 448 free_pages_exact(*vaddr, nr_pages * XEN_PAGE_SIZE); in xenbus_teardown_ring() 749 area = get_vm_area(XEN_PAGE_SIZE * nr_grefs, VM_IOREMAP); in xenbus_map_ring_pv() 753 XEN_PAGE_SIZE * nr_grefs, map_ring_apply, info)) in xenbus_map_ring_pv() 812 addr = (unsigned long)vaddr + (XEN_PAGE_SIZE * i); in xenbus_unmap_ring_pv()
|
D | xenbus_probe.c | 756 XEN_PAGE_SIZE, MEMREMAP_WB); in xenbus_probe() 1013 XEN_PAGE_SIZE, MEMREMAP_WB); in xenbus_init()
|
/linux-5.19.10/drivers/block/xen-blkback/ |
D | xenbus.c | 221 rsp_prod, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map() 222 size = __RING_SIZE(sring_native, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map() 231 rsp_prod, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map() 232 size = __RING_SIZE(sring_x86_32, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map() 241 rsp_prod, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map() 242 size = __RING_SIZE(sring_x86_64, XEN_PAGE_SIZE * nr_grefs); in xen_blkif_map()
|
D | common.h | 59 #define XEN_PAGES_PER_SEGMENT (PAGE_SIZE / XEN_PAGE_SIZE) 62 (XEN_PAGE_SIZE/sizeof(struct blkif_request_segment))
|
/linux-5.19.10/sound/xen/ |
D | xen_snd_front.c | 373 if (XEN_PAGE_SIZE != PAGE_SIZE) { in xen_drv_init() 375 XEN_PAGE_SIZE, PAGE_SIZE); in xen_drv_init()
|
/linux-5.19.10/include/xen/interface/io/ |
D | displif.h | 921 #define XENDISPL_EVENT_PAGE_SIZE XEN_PAGE_SIZE
|
D | sndif.h | 1055 #define XENSND_EVENT_PAGE_SIZE XEN_PAGE_SIZE
|
/linux-5.19.10/drivers/net/ |
D | xen-netfront.c | 85 #define NET_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, XEN_PAGE_SIZE) 86 #define NET_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, XEN_PAGE_SIZE) 681 XEN_PAGE_SIZE); in bounce_skb() 687 if (!IS_ALIGNED((uintptr_t)n->head, XEN_PAGE_SIZE)) { in bounce_skb() 704 #define MAX_XEN_SKB_FRAGS (65536 / XEN_PAGE_SIZE + 1) 976 xdp_init_buff(xdp, XEN_PAGE_SIZE - XDP_PACKET_HEADROOM, in xennet_run_xdp() 1047 rx->offset + rx->status > XEN_PAGE_SIZE)) { in xennet_get_responses() 1620 unsigned long max_mtu = XEN_PAGE_SIZE - XDP_PACKET_HEADROOM; in xennet_xdp_set() 1977 XEN_FRONT_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE); in setup_netfront() 1984 XEN_FRONT_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE); in setup_netfront() [all …]
|
/linux-5.19.10/drivers/tty/hvc/ |
D | hvc_xen.c | 256 info->intf = memremap(gfn << XEN_PAGE_SHIFT, XEN_PAGE_SIZE, MEMREMAP_WB); in xen_hvm_console_init() 505 memset(info->intf, 0, XEN_PAGE_SIZE); in xencons_resume()
|
/linux-5.19.10/drivers/gpu/drm/xen/ |
D | xen_drm_front.c | 773 if (XEN_PAGE_SIZE != PAGE_SIZE) { in xen_drv_init() 775 XEN_PAGE_SIZE, PAGE_SIZE); in xen_drv_init()
|
/linux-5.19.10/drivers/block/ |
D | xen-blkfront.c | 160 __CONST_RING_SIZE(blkif, XEN_PAGE_SIZE * (info)->nr_ring_pages) 262 #define GRANTS_PER_PSEG (PAGE_SIZE / XEN_PAGE_SIZE) 265 (XEN_PAGE_SIZE / sizeof(struct blkif_request_segment)) 696 (BLKIF_MAX_SEGMENTS_PER_REQUEST * XEN_PAGE_SIZE) / 512; in blkif_setup_extra_req() 965 blk_queue_max_hw_sectors(rq, (segments * XEN_PAGE_SIZE) / 512); in blkif_set_queue_limits() 1686 unsigned long ring_size = info->nr_ring_pages * XEN_PAGE_SIZE; in setup_blkring()
|
/linux-5.19.10/drivers/xen/events/ |
D | events_fifo.c | 57 #define EVENT_WORDS_PER_PAGE (XEN_PAGE_SIZE / sizeof(event_word_t))
|