Searched refs:eoi_mmio (Results 1 – 6 of 6) sorted by relevance
428 data->eoi_mmio = ioremap(data->eoi_page, 1u << data->esb_shift); in xive_spapr_populate_irq_data()429 if (!data->eoi_mmio) { in xive_spapr_populate_irq_data()436 data->trig_mmio = data->eoi_mmio; in xive_spapr_populate_irq_data()442 iounmap(data->eoi_mmio); in xive_spapr_populate_irq_data()
227 val = in_be64(xd->eoi_mmio + offset); in xive_esb_read()237 out_be64(xd->eoi_mmio + offset, data); in xive_esb_write()1001 if (xd->eoi_mmio) { in xive_cleanup_irq_data()1002 iounmap(xd->eoi_mmio); in xive_cleanup_irq_data()1003 if (xd->eoi_mmio == xd->trig_mmio) in xive_cleanup_irq_data()1005 xd->eoi_mmio = NULL; in xive_cleanup_irq_data()
75 data->eoi_mmio = ioremap(data->eoi_page, 1u << data->esb_shift); in xive_native_populate_irq_data()76 if (!data->eoi_mmio) { in xive_native_populate_irq_data()86 data->trig_mmio = data->eoi_mmio; in xive_native_populate_irq_data()
42 void __iomem *eoi_mmio; member
33 #define __x_eoi_page(xd) ((void __iomem *)((xd)->eoi_mmio))924 vcpu->arch.xive_esc_vaddr = (__force u64)xd->eoi_mmio; in kvmppc_xive_attach_escalation()
40 val = in_be64(xd->eoi_mmio + offset); in xive_vm_esb_load()