Searched refs:dma_need_sync (Results 1 – 7 of 7) sorted by relevance
38 bool dma_need_sync; member75 bool dma_need_sync; member157 if (!pool->dma_need_sync) in xp_dma_sync_for_device()
127 if (!pool->dma_need_sync) in xsk_buff_dma_sync_for_cpu()
313 dma_map->dma_need_sync = false; in xp_create_dma_map()398 pool->dma_need_sync = dma_map->dma_need_sync; in xp_init_dma_info()434 if (dma_need_sync(dev, dma)) in xp_dma_map()435 dma_map->dma_need_sync = true; in xp_dma_map()530 if (pool->dma_need_sync) { in xp_alloc()606 if (unlikely(pool->dma_need_sync)) { in xp_alloc_batch()
147 bool dma_need_sync(struct device *dev, dma_addr_t dma_addr);269 static inline bool dma_need_sync(struct device *dev, dma_addr_t dma_addr) in dma_need_sync() function
776 bool dma_need_sync(struct device *dev, dma_addr_t dma_addr) in dma_need_sync() function784 EXPORT_SYMBOL_GPL(dma_need_sync);
535 dmab->dev.need_sync = dma_need_sync(dmab->dev.dev, in snd_dma_noncontig_alloc()824 dmab->dev.need_sync = dma_need_sync(dmab->dev.dev, dmab->addr); in snd_dma_noncoherent_alloc()
210 dma_need_sync(struct device *dev, dma_addr_t dma_addr);